ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k8_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8501
  • Qwk: 0.2726
  • Mse: 0.8501
  • Rmse: 0.9220

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1053 2 2.4641 -0.0568 2.4641 1.5697
No log 0.2105 4 1.2791 0.1882 1.2791 1.1310
No log 0.3158 6 1.0031 -0.0550 1.0031 1.0015
No log 0.4211 8 1.1737 -0.1355 1.1737 1.0834
No log 0.5263 10 1.2532 -0.1993 1.2532 1.1195
No log 0.6316 12 0.8472 0.0 0.8472 0.9204
No log 0.7368 14 0.6607 0.1232 0.6607 0.8128
No log 0.8421 16 0.6433 0.2676 0.6433 0.8021
No log 0.9474 18 0.7040 0.3019 0.7040 0.8391
No log 1.0526 20 0.7188 0.3019 0.7188 0.8478
No log 1.1579 22 0.6808 0.2676 0.6808 0.8251
No log 1.2632 24 0.6844 0.2676 0.6844 0.8273
No log 1.3684 26 0.6940 0.2676 0.6940 0.8331
No log 1.4737 28 0.7603 0.3125 0.7603 0.8720
No log 1.5789 30 0.8185 0.1660 0.8185 0.9047
No log 1.6842 32 0.7336 0.2748 0.7336 0.8565
No log 1.7895 34 0.8494 0.2358 0.8494 0.9217
No log 1.8947 36 1.0127 0.1955 1.0127 1.0063
No log 2.0 38 0.8270 0.1459 0.8270 0.9094
No log 2.1053 40 0.7232 0.1277 0.7232 0.8504
No log 2.2105 42 0.7874 0.2156 0.7874 0.8874
No log 2.3158 44 0.7714 0.1365 0.7714 0.8783
No log 2.4211 46 0.7528 0.0717 0.7528 0.8677
No log 2.5263 48 0.8233 0.2407 0.8233 0.9073
No log 2.6316 50 0.8316 0.2652 0.8316 0.9119
No log 2.7368 52 0.7702 0.1863 0.7702 0.8776
No log 2.8421 54 0.7735 0.2884 0.7735 0.8795
No log 2.9474 56 0.7935 0.3238 0.7935 0.8908
No log 3.0526 58 1.0933 0.1241 1.0933 1.0456
No log 3.1579 60 1.2316 0.1839 1.2316 1.1098
No log 3.2632 62 1.0365 0.2460 1.0365 1.0181
No log 3.3684 64 0.9874 0.1692 0.9874 0.9937
No log 3.4737 66 1.1109 0.2209 1.1109 1.0540
No log 3.5789 68 1.4546 0.1067 1.4546 1.2061
No log 3.6842 70 1.6366 0.1555 1.6366 1.2793
No log 3.7895 72 1.3934 0.1093 1.3934 1.1804
No log 3.8947 74 1.3039 0.1175 1.3039 1.1419
No log 4.0 76 1.1431 0.1394 1.1431 1.0692
No log 4.1053 78 1.1404 0.1976 1.1404 1.0679
No log 4.2105 80 1.3409 0.1568 1.3409 1.1580
No log 4.3158 82 1.1559 0.1618 1.1559 1.0751
No log 4.4211 84 1.2002 0.1799 1.2002 1.0955
No log 4.5263 86 1.5612 0.1169 1.5612 1.2495
No log 4.6316 88 1.9707 0.0421 1.9707 1.4038
No log 4.7368 90 1.9777 0.0421 1.9777 1.4063
No log 4.8421 92 1.7506 0.0589 1.7506 1.3231
No log 4.9474 94 1.5731 0.1195 1.5731 1.2542
No log 5.0526 96 1.4243 0.1093 1.4243 1.1934
No log 5.1579 98 1.3726 0.1093 1.3726 1.1716
No log 5.2632 100 1.6732 0.0689 1.6732 1.2935
No log 5.3684 102 1.6295 0.0300 1.6295 1.2765
No log 5.4737 104 1.1847 0.1029 1.1847 1.0885
No log 5.5789 106 0.9579 0.1661 0.9579 0.9787
No log 5.6842 108 0.9987 0.1603 0.9987 0.9994
No log 5.7895 110 1.2834 0.1458 1.2834 1.1329
No log 5.8947 112 1.4731 0.0803 1.4731 1.2137
No log 6.0 114 1.3215 0.1458 1.3215 1.1496
No log 6.1053 116 0.9663 0.1651 0.9663 0.9830
No log 6.2105 118 0.9060 0.2076 0.9060 0.9518
No log 6.3158 120 1.0266 0.1210 1.0266 1.0132
No log 6.4211 122 1.4518 0.0361 1.4518 1.2049
No log 6.5263 124 1.8053 0.0350 1.8053 1.3436
No log 6.6316 126 1.7976 0.0350 1.7976 1.3408
No log 6.7368 128 1.5457 0.0447 1.5457 1.2432
No log 6.8421 130 1.2204 0.1262 1.2204 1.1047
No log 6.9474 132 1.1004 0.1356 1.1004 1.0490
No log 7.0526 134 1.1797 0.1293 1.1797 1.0861
No log 7.1579 136 1.3349 0.1174 1.3349 1.1554
No log 7.2632 138 1.2656 0.1293 1.2656 1.1250
No log 7.3684 140 1.2029 0.1293 1.2029 1.0968
No log 7.4737 142 1.3612 0.1175 1.3612 1.1667
No log 7.5789 144 1.5873 0.0283 1.5873 1.2599
No log 7.6842 146 1.5825 0.0283 1.5825 1.2580
No log 7.7895 148 1.3211 0.1175 1.3211 1.1494
No log 7.8947 150 1.0630 0.2119 1.0630 1.0310
No log 8.0 152 1.0602 0.1787 1.0602 1.0297
No log 8.1053 154 1.1823 0.2412 1.1823 1.0874
No log 8.2105 156 1.3901 0.0873 1.3901 1.1790
No log 8.3158 158 1.3194 0.1464 1.3194 1.1486
No log 8.4211 160 1.0929 0.1709 1.0929 1.0454
No log 8.5263 162 0.9093 0.1777 0.9093 0.9536
No log 8.6316 164 0.8668 0.2692 0.8668 0.9310
No log 8.7368 166 0.9346 0.1955 0.9346 0.9667
No log 8.8421 168 1.2601 0.1458 1.2601 1.1225
No log 8.9474 170 1.4888 0.0745 1.4888 1.2202
No log 9.0526 172 1.3814 0.0829 1.3814 1.1753
No log 9.1579 174 1.0769 0.2782 1.0769 1.0378
No log 9.2632 176 0.8222 0.2352 0.8222 0.9067
No log 9.3684 178 0.7644 0.2407 0.7644 0.8743
No log 9.4737 180 0.7745 0.2718 0.7745 0.8800
No log 9.5789 182 0.9117 0.2000 0.9117 0.9549
No log 9.6842 184 1.2036 0.2045 1.2036 1.0971
No log 9.7895 186 1.3599 0.1427 1.3599 1.1662
No log 9.8947 188 1.3856 0.1275 1.3856 1.1771
No log 10.0 190 1.4587 0.1549 1.4587 1.2078
No log 10.1053 192 1.2876 0.1638 1.2876 1.1347
No log 10.2105 194 1.1956 0.1784 1.1956 1.0934
No log 10.3158 196 1.1869 0.1784 1.1869 1.0894
No log 10.4211 198 1.2118 0.1490 1.2118 1.1008
No log 10.5263 200 1.2318 0.1458 1.2318 1.1099
No log 10.6316 202 1.2338 0.1458 1.2338 1.1108
No log 10.7368 204 1.1210 0.1626 1.1210 1.0588
No log 10.8421 206 0.9735 0.1274 0.9735 0.9867
No log 10.9474 208 0.9787 0.1557 0.9787 0.9893
No log 11.0526 210 1.0502 0.0925 1.0502 1.0248
No log 11.1579 212 1.2232 0.1943 1.2232 1.1060
No log 11.2632 214 1.3664 0.1220 1.3664 1.1689
No log 11.3684 216 1.2836 0.1427 1.2836 1.1329
No log 11.4737 218 1.1297 0.2782 1.1297 1.0629
No log 11.5789 220 1.0821 0.2032 1.0821 1.0402
No log 11.6842 222 1.0552 0.1787 1.0552 1.0272
No log 11.7895 224 1.1177 0.1949 1.1177 1.0572
No log 11.8947 226 1.2791 0.0712 1.2791 1.1310
No log 12.0 228 1.3577 0.0419 1.3577 1.1652
No log 12.1053 230 1.3338 0.0694 1.3338 1.1549
No log 12.2105 232 1.5370 0.0832 1.5370 1.2398
No log 12.3158 234 1.7838 0.0932 1.7838 1.3356
No log 12.4211 236 1.6514 0.1549 1.6514 1.2851
No log 12.5263 238 1.3468 0.0952 1.3468 1.1605
No log 12.6316 240 1.1144 0.0448 1.1144 1.0557
No log 12.7368 242 1.0226 0.0448 1.0226 1.0112
No log 12.8421 244 1.0179 0.0799 1.0179 1.0089
No log 12.9474 246 1.1405 0.0585 1.1405 1.0679
No log 13.0526 248 1.2102 0.1205 1.2102 1.1001
No log 13.1579 250 1.1881 0.0761 1.1881 1.0900
No log 13.2632 252 1.0965 0.0982 1.0965 1.0472
No log 13.3684 254 1.1490 0.1243 1.1490 1.0719
No log 13.4737 256 1.3393 0.0694 1.3393 1.1573
No log 13.5789 258 1.5029 0.0086 1.5029 1.2259
No log 13.6842 260 1.4680 0.0086 1.4680 1.2116
No log 13.7895 262 1.2850 0.0459 1.2850 1.1336
No log 13.8947 264 1.2543 0.0761 1.2543 1.1200
No log 14.0 266 1.1445 0.1147 1.1445 1.0698
No log 14.1053 268 1.0257 0.1312 1.0257 1.0128
No log 14.2105 270 1.0147 0.1734 1.0147 1.0073
No log 14.3158 272 1.1565 0.0561 1.1565 1.0754
No log 14.4211 274 1.4045 0.0584 1.4045 1.1851
No log 14.5263 276 1.5301 0.0230 1.5301 1.2370
No log 14.6316 278 1.4397 0.0531 1.4397 1.1999
No log 14.7368 280 1.3048 0.1458 1.3048 1.1423
No log 14.8421 282 1.2171 0.0790 1.2171 1.1032
No log 14.9474 284 1.1725 0.0546 1.1725 1.0828
No log 15.0526 286 1.2450 0.1262 1.2450 1.1158
No log 15.1579 288 1.3341 0.0627 1.3341 1.1550
No log 15.2632 290 1.3896 0.0895 1.3896 1.1788
No log 15.3684 292 1.2685 0.0921 1.2685 1.1263
No log 15.4737 294 1.1053 0.1832 1.1053 1.0514
No log 15.5789 296 1.0107 0.2032 1.0107 1.0053
No log 15.6842 298 0.9490 0.1651 0.9490 0.9741
No log 15.7895 300 0.9145 0.1822 0.9145 0.9563
No log 15.8947 302 0.9693 0.0982 0.9693 0.9846
No log 16.0 304 1.0902 0.1747 1.0902 1.0441
No log 16.1053 306 1.2888 0.1203 1.2888 1.1353
No log 16.2105 308 1.5212 0.0519 1.5212 1.2334
No log 16.3158 310 1.5173 0.0519 1.5173 1.2318
No log 16.4211 312 1.3072 0.1233 1.3072 1.1433
No log 16.5263 314 1.1155 0.1086 1.1155 1.0562
No log 16.6316 316 1.0768 0.0569 1.0768 1.0377
No log 16.7368 318 1.0070 0.0982 1.0070 1.0035
No log 16.8421 320 1.0212 0.1210 1.0212 1.0105
No log 16.9474 322 1.0584 0.1178 1.0584 1.0288
No log 17.0526 324 1.1352 0.1328 1.1352 1.0654
No log 17.1579 326 1.0882 0.1949 1.0882 1.0432
No log 17.2632 328 1.0083 0.1210 1.0083 1.0041
No log 17.3684 330 1.0760 0.1635 1.0760 1.0373
No log 17.4737 332 1.0971 0.1530 1.0971 1.0474
No log 17.5789 334 1.1515 0.1530 1.1515 1.0731
No log 17.6842 336 1.1834 0.1530 1.1834 1.0879
No log 17.7895 338 1.1875 0.1870 1.1875 1.0897
No log 17.8947 340 1.2742 0.1203 1.2742 1.1288
No log 18.0 342 1.1476 0.1821 1.1476 1.0713
No log 18.1053 344 0.9655 0.2017 0.9655 0.9826
No log 18.2105 346 0.8719 0.2949 0.8719 0.9337
No log 18.3158 348 0.8436 0.2754 0.8436 0.9185
No log 18.4211 350 0.8642 0.2358 0.8642 0.9296
No log 18.5263 352 0.9297 0.2211 0.9297 0.9642
No log 18.6316 354 1.1023 0.1990 1.1023 1.0499
No log 18.7368 356 1.2024 0.2183 1.2024 1.0966
No log 18.8421 358 1.2440 0.1490 1.2440 1.1154
No log 18.9474 360 1.2552 0.1490 1.2552 1.1204
No log 19.0526 362 1.1656 0.2227 1.1656 1.0796
No log 19.1579 364 1.0296 0.1210 1.0296 1.0147
No log 19.2632 366 1.0175 0.1535 1.0175 1.0087
No log 19.3684 368 1.0279 0.1535 1.0279 1.0138
No log 19.4737 370 1.1408 0.1870 1.1408 1.0681
No log 19.5789 372 1.2388 0.1464 1.2388 1.1130
No log 19.6842 374 1.2974 0.1176 1.2974 1.1390
No log 19.7895 376 1.2912 0.1205 1.2912 1.1363
No log 19.8947 378 1.3419 0.1205 1.3419 1.1584
No log 20.0 380 1.2683 0.1233 1.2683 1.1262
No log 20.1053 382 1.1439 0.1115 1.1439 1.0695
No log 20.2105 384 1.0252 0.1277 1.0252 1.0125
No log 20.3158 386 1.0168 0.1243 1.0168 1.0083
No log 20.4211 388 1.1033 0.1210 1.1033 1.0504
No log 20.5263 390 1.2257 0.2227 1.2257 1.1071
No log 20.6316 392 1.2128 0.2183 1.2128 1.1013
No log 20.7368 394 1.1028 0.1990 1.1028 1.0501
No log 20.8421 396 0.9732 0.1612 0.9732 0.9865
No log 20.9474 398 0.8972 0.2046 0.8972 0.9472
No log 21.0526 400 0.8210 0.2467 0.8210 0.9061
No log 21.1579 402 0.8382 0.1962 0.8382 0.9155
No log 21.2632 404 0.9694 0.0686 0.9694 0.9846
No log 21.3684 406 1.2246 0.1463 1.2246 1.1066
No log 21.4737 408 1.4683 0.1272 1.4683 1.2117
No log 21.5789 410 1.6180 0.0674 1.6180 1.2720
No log 21.6842 412 1.6021 0.0905 1.6021 1.2658
No log 21.7895 414 1.5189 0.0724 1.5189 1.2325
No log 21.8947 416 1.3595 0.1530 1.3595 1.1660
No log 22.0 418 1.2128 0.1671 1.2128 1.1013
No log 22.1053 420 1.1519 0.1709 1.1519 1.0733
No log 22.2105 422 1.0861 0.1463 1.0861 1.0422
No log 22.3158 424 1.0054 0.0952 1.0054 1.0027
No log 22.4211 426 0.9362 0.1461 0.9362 0.9676
No log 22.5263 428 0.9989 0.1277 0.9989 0.9994
No log 22.6316 430 0.9631 0.1348 0.9631 0.9814
No log 22.7368 432 0.9612 0.1348 0.9612 0.9804
No log 22.8421 434 0.9709 0.1312 0.9709 0.9853
No log 22.9474 436 0.9820 0.1312 0.9820 0.9910
No log 23.0526 438 0.9666 0.1312 0.9666 0.9832
No log 23.1579 440 0.9325 0.1312 0.9325 0.9656
No log 23.2632 442 0.9306 0.1312 0.9306 0.9647
No log 23.3684 444 0.9622 0.1573 0.9622 0.9809
No log 23.4737 446 0.9685 0.2119 0.9685 0.9841
No log 23.5789 448 0.9311 0.2259 0.9311 0.9650
No log 23.6842 450 0.9567 0.2810 0.9567 0.9781
No log 23.7895 452 1.0083 0.2316 1.0083 1.0041
No log 23.8947 454 1.0825 0.2504 1.0825 1.0404
No log 24.0 456 1.0452 0.2552 1.0452 1.0224
No log 24.1053 458 1.0059 0.1787 1.0059 1.0029
No log 24.2105 460 0.9396 0.2211 0.9396 0.9693
No log 24.3158 462 0.9184 0.2211 0.9184 0.9583
No log 24.4211 464 0.9420 0.2211 0.9420 0.9706
No log 24.5263 466 1.0272 0.2552 1.0272 1.0135
No log 24.6316 468 1.1672 0.2412 1.1672 1.0804
No log 24.7368 470 1.2095 0.1748 1.2095 1.0998
No log 24.8421 472 1.1493 0.2166 1.1493 1.0721
No log 24.9474 474 1.0175 0.2552 1.0175 1.0087
No log 25.0526 476 0.8896 0.2615 0.8896 0.9432
No log 25.1579 478 0.8071 0.2784 0.8071 0.8984
No log 25.2632 480 0.7829 0.3234 0.7829 0.8848
No log 25.3684 482 0.7967 0.2726 0.7967 0.8926
No log 25.4737 484 0.8690 0.2615 0.8690 0.9322
No log 25.5789 486 0.9307 0.3643 0.9307 0.9647
No log 25.6842 488 0.9252 0.3643 0.9252 0.9619
No log 25.7895 490 0.9541 0.3579 0.9541 0.9768
No log 25.8947 492 0.9794 0.3228 0.9794 0.9896
No log 26.0 494 0.9896 0.2363 0.9896 0.9948
No log 26.1053 496 0.9339 0.1869 0.9339 0.9664
No log 26.2105 498 0.8779 0.2670 0.8779 0.9369
0.2427 26.3158 500 0.8182 0.2784 0.8182 0.9045
0.2427 26.4211 502 0.8560 0.2726 0.8560 0.9252
0.2427 26.5263 504 0.9091 0.2211 0.9091 0.9535
0.2427 26.6316 506 0.8502 0.2726 0.8502 0.9221
0.2427 26.7368 508 0.8398 0.2726 0.8398 0.9164
0.2427 26.8421 510 0.8501 0.2726 0.8501 0.9220

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k8_task7_organization

Finetuned
(4019)
this model