ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k7_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.3044
  • Qwk: 0.0401
  • Mse: 1.3044
  • Rmse: 1.1421

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1176 2 4.0140 -0.0174 4.0140 2.0035
No log 0.2353 4 2.6063 -0.0926 2.6063 1.6144
No log 0.3529 6 1.5650 -0.0684 1.5650 1.2510
No log 0.4706 8 1.1358 0.1160 1.1358 1.0658
No log 0.5882 10 1.0366 0.2541 1.0366 1.0181
No log 0.7059 12 1.1396 0.2515 1.1396 1.0675
No log 0.8235 14 1.1764 0.1240 1.1764 1.0846
No log 0.9412 16 1.1849 0.1148 1.1849 1.0885
No log 1.0588 18 1.0913 0.2243 1.0913 1.0447
No log 1.1765 20 1.0292 0.1799 1.0292 1.0145
No log 1.2941 22 1.1401 0.1755 1.1401 1.0678
No log 1.4118 24 1.3986 -0.1078 1.3986 1.1826
No log 1.5294 26 1.4496 0.0 1.4496 1.2040
No log 1.6471 28 1.2535 0.0380 1.2535 1.1196
No log 1.7647 30 1.2141 0.0996 1.2141 1.1019
No log 1.8824 32 1.1102 0.2221 1.1102 1.0536
No log 2.0 34 1.0261 0.2692 1.0261 1.0130
No log 2.1176 36 1.0550 0.2569 1.0550 1.0271
No log 2.2353 38 1.0972 0.1545 1.0972 1.0475
No log 2.3529 40 1.0478 0.2314 1.0478 1.0236
No log 2.4706 42 1.0741 0.1981 1.0741 1.0364
No log 2.5882 44 1.0318 0.2492 1.0318 1.0158
No log 2.7059 46 1.0765 0.1601 1.0765 1.0376
No log 2.8235 48 1.5068 0.1081 1.5068 1.2275
No log 2.9412 50 1.6387 0.0673 1.6387 1.2801
No log 3.0588 52 1.4641 0.1230 1.4641 1.2100
No log 3.1765 54 1.1123 0.1830 1.1123 1.0546
No log 3.2941 56 1.0415 0.3134 1.0415 1.0206
No log 3.4118 58 1.0282 0.3011 1.0282 1.0140
No log 3.5294 60 1.1434 0.1530 1.1434 1.0693
No log 3.6471 62 1.7029 -0.1001 1.7029 1.3049
No log 3.7647 64 1.8308 -0.1806 1.8308 1.3531
No log 3.8824 66 1.6226 -0.2240 1.6226 1.2738
No log 4.0 68 1.2734 0.0169 1.2734 1.1285
No log 4.1176 70 0.9954 0.2547 0.9954 0.9977
No log 4.2353 72 0.9919 0.1734 0.9919 0.9959
No log 4.3529 74 1.1816 0.1707 1.1816 1.0870
No log 4.4706 76 1.6376 -0.1434 1.6376 1.2797
No log 4.5882 78 1.6640 -0.0513 1.6640 1.2899
No log 4.7059 80 1.5073 0.1292 1.5073 1.2277
No log 4.8235 82 1.5057 0.1477 1.5057 1.2271
No log 4.9412 84 1.6460 0.1537 1.6460 1.2830
No log 5.0588 86 1.8272 0.1379 1.8272 1.3517
No log 5.1765 88 1.8576 0.1367 1.8576 1.3629
No log 5.2941 90 1.6237 0.2163 1.6237 1.2742
No log 5.4118 92 1.5985 0.2026 1.5985 1.2643
No log 5.5294 94 1.7968 0.1771 1.7968 1.3405
No log 5.6471 96 1.9635 -0.0200 1.9635 1.4013
No log 5.7647 98 1.9162 -0.0766 1.9162 1.3843
No log 5.8824 100 1.7822 0.0357 1.7822 1.3350
No log 6.0 102 1.6227 0.1656 1.6227 1.2739
No log 6.1176 104 1.6403 0.0969 1.6403 1.2807
No log 6.2353 106 1.7658 -0.0009 1.7658 1.3288
No log 6.3529 108 1.7952 0.0181 1.7952 1.3399
No log 6.4706 110 1.7833 0.0749 1.7833 1.3354
No log 6.5882 112 1.7027 0.0749 1.7027 1.3049
No log 6.7059 114 1.6791 0.0749 1.6791 1.2958
No log 6.8235 116 1.5693 0.1193 1.5693 1.2527
No log 6.9412 118 1.5476 0.0760 1.5476 1.2440
No log 7.0588 120 1.4455 0.1255 1.4455 1.2023
No log 7.1765 122 1.4469 0.0907 1.4469 1.2029
No log 7.2941 124 1.6220 0.1323 1.6220 1.2736
No log 7.4118 126 1.6811 0.0829 1.6811 1.2966
No log 7.5294 128 1.5359 0.2016 1.5359 1.2393
No log 7.6471 130 1.1754 0.2707 1.1754 1.0841
No log 7.7647 132 1.1192 0.2906 1.1192 1.0579
No log 7.8824 134 1.3121 0.2367 1.3121 1.1455
No log 8.0 136 1.5358 0.2006 1.5358 1.2393
No log 8.1176 138 1.4863 0.2317 1.4863 1.2191
No log 8.2353 140 1.3648 0.2424 1.3648 1.1682
No log 8.3529 142 1.5123 0.2058 1.5123 1.2297
No log 8.4706 144 1.5114 0.2270 1.5114 1.2294
No log 8.5882 146 1.3171 0.1562 1.3171 1.1476
No log 8.7059 148 1.3334 0.1486 1.3334 1.1547
No log 8.8235 150 1.4493 0.1814 1.4493 1.2039
No log 8.9412 152 1.5059 0.2292 1.5059 1.2272
No log 9.0588 154 1.3437 0.1486 1.3437 1.1592
No log 9.1765 156 1.1978 0.0931 1.1978 1.0945
No log 9.2941 158 1.1843 0.0931 1.1843 1.0882
No log 9.4118 160 1.2939 0.1486 1.2939 1.1375
No log 9.5294 162 1.5258 0.1955 1.5258 1.2352
No log 9.6471 164 1.4207 0.2342 1.4207 1.1919
No log 9.7647 166 1.1949 0.1628 1.1949 1.0931
No log 9.8824 168 1.1262 0.1649 1.1262 1.0612
No log 10.0 170 1.1602 0.1202 1.1602 1.0771
No log 10.1176 172 1.1975 0.1202 1.1975 1.0943
No log 10.2353 174 1.5264 0.2110 1.5264 1.2355
No log 10.3529 176 1.6088 0.1688 1.6088 1.2684
No log 10.4706 178 1.4086 0.1630 1.4086 1.1868
No log 10.5882 180 1.2384 0.1579 1.2384 1.1129
No log 10.7059 182 1.2859 0.1228 1.2859 1.1340
No log 10.8235 184 1.5012 0.2437 1.5012 1.2252
No log 10.9412 186 1.5139 0.2694 1.5139 1.2304
No log 11.0588 188 1.4150 0.2062 1.4150 1.1895
No log 11.1765 190 1.2263 0.0781 1.2263 1.1074
No log 11.2941 192 1.1721 0.0401 1.1721 1.0826
No log 11.4118 194 1.2260 0.0401 1.2260 1.1072
No log 11.5294 196 1.3175 0.1052 1.3175 1.1478
No log 11.6471 198 1.4581 0.1832 1.4581 1.2075
No log 11.7647 200 1.6562 0.1635 1.6562 1.2869
No log 11.8824 202 1.6667 0.1548 1.6667 1.2910
No log 12.0 204 1.6903 0.0733 1.6903 1.3001
No log 12.1176 206 1.7801 0.1075 1.7801 1.3342
No log 12.2353 208 1.6606 0.1729 1.6606 1.2887
No log 12.3529 210 1.4356 0.1310 1.4356 1.1982
No log 12.4706 212 1.2895 0.0401 1.2895 1.1356
No log 12.5882 214 1.2159 0.0556 1.2159 1.1027
No log 12.7059 216 1.3015 0.1407 1.3015 1.1409
No log 12.8235 218 1.4769 0.1769 1.4769 1.2153
No log 12.9412 220 1.6701 0.1344 1.6701 1.2923
No log 13.0588 222 1.7238 0.2005 1.7238 1.3129
No log 13.1765 224 1.6111 0.2062 1.6111 1.2693
No log 13.2941 226 1.4191 0.1562 1.4191 1.1913
No log 13.4118 228 1.3272 0.1202 1.3272 1.1521
No log 13.5294 230 1.3933 0.1700 1.3933 1.1804
No log 13.6471 232 1.4272 0.1769 1.4272 1.1947
No log 13.7647 234 1.4810 0.1892 1.4810 1.2170
No log 13.8824 236 1.3777 0.1703 1.3777 1.1737
No log 14.0 238 1.3419 0.2004 1.3419 1.1584
No log 14.1176 240 1.2781 0.1835 1.2781 1.1305
No log 14.2353 242 1.3544 0.2004 1.3544 1.1638
No log 14.3529 244 1.5277 0.2221 1.5277 1.2360
No log 14.4706 246 1.6322 0.2317 1.6322 1.2776
No log 14.5882 248 1.5557 0.2062 1.5557 1.2473
No log 14.7059 250 1.4156 0.1228 1.4156 1.1898
No log 14.8235 252 1.2713 0.0556 1.2713 1.1275
No log 14.9412 254 1.2162 0.0710 1.2162 1.1028
No log 15.0588 256 1.2566 0.0556 1.2566 1.1210
No log 15.1765 258 1.3755 0.1052 1.3755 1.1728
No log 15.2941 260 1.3939 0.0510 1.3939 1.1806
No log 15.4118 262 1.4036 0.1142 1.4036 1.1847
No log 15.5294 264 1.5289 0.1943 1.5289 1.2365
No log 15.6471 266 1.5388 0.2482 1.5388 1.2405
No log 15.7647 268 1.6170 0.1807 1.6170 1.2716
No log 15.8824 270 1.7945 0.0892 1.7945 1.3396
No log 16.0 272 1.7848 0.1344 1.7848 1.3360
No log 16.1176 274 1.5707 0.1562 1.5707 1.2533
No log 16.2353 276 1.4000 0.0510 1.4000 1.1832
No log 16.3529 278 1.3294 0.0401 1.3294 1.1530
No log 16.4706 280 1.2594 0.0401 1.2594 1.1222
No log 16.5882 282 1.2913 0.1202 1.2913 1.1363
No log 16.7059 284 1.4278 0.1228 1.4278 1.1949
No log 16.8235 286 1.6698 0.2058 1.6698 1.2922
No log 16.9412 288 1.7508 0.1525 1.7508 1.3232
No log 17.0588 290 1.6792 0.2110 1.6792 1.2959
No log 17.1765 292 1.5453 0.2437 1.5453 1.2431
No log 17.2941 294 1.4507 0.2126 1.4507 1.2044
No log 17.4118 296 1.3099 0.0 1.3099 1.1445
No log 17.5294 298 1.2884 0.0 1.2884 1.1351
No log 17.6471 300 1.2905 0.0 1.2905 1.1360
No log 17.7647 302 1.2815 0.0 1.2815 1.1320
No log 17.8824 304 1.3225 0.0 1.3225 1.1500
No log 18.0 306 1.4414 0.2126 1.4414 1.2006
No log 18.1176 308 1.6130 0.2058 1.6130 1.2700
No log 18.2353 310 1.6717 0.1140 1.6717 1.2929
No log 18.3529 312 1.8210 0.1323 1.8210 1.3495
No log 18.4706 314 1.7203 0.1323 1.7203 1.3116
No log 18.5882 316 1.4580 0.2391 1.4580 1.2075
No log 18.7059 318 1.3334 0.1024 1.3334 1.1547
No log 18.8235 320 1.4236 0.1634 1.4236 1.1931
No log 18.9412 322 1.4935 0.2474 1.4935 1.2221
No log 19.0588 324 1.5170 0.2474 1.5170 1.2316
No log 19.1765 326 1.5111 0.2065 1.5111 1.2292
No log 19.2941 328 1.5045 0.2126 1.5045 1.2266
No log 19.4118 330 1.5585 0.2126 1.5585 1.2484
No log 19.5294 332 1.5062 0.1486 1.5062 1.2273
No log 19.6471 334 1.4935 0.1142 1.4935 1.2221
No log 19.7647 336 1.4867 0.0878 1.4867 1.2193
No log 19.8824 338 1.4005 0.0510 1.4005 1.1834
No log 20.0 340 1.3361 0.0401 1.3361 1.1559
No log 20.1176 342 1.3090 0.0401 1.3090 1.1441
No log 20.2353 344 1.4213 0.1744 1.4213 1.1922
No log 20.3529 346 1.4968 0.2731 1.4968 1.2234
No log 20.4706 348 1.6010 0.2518 1.6010 1.2653
No log 20.5882 350 1.7172 0.1893 1.7172 1.3104
No log 20.7059 352 1.8887 0.1772 1.8887 1.3743
No log 20.8235 354 1.9277 0.1815 1.9277 1.3884
No log 20.9412 356 1.8766 0.1880 1.8766 1.3699
No log 21.0588 358 1.6474 0.1955 1.6474 1.2835
No log 21.1765 360 1.4019 0.2015 1.4019 1.1840
No log 21.2941 362 1.3020 0.0931 1.3020 1.1411
No log 21.4118 364 1.4082 0.2752 1.4082 1.1867
No log 21.5294 366 1.6080 0.2644 1.6080 1.2681
No log 21.6471 368 1.6065 0.2806 1.6065 1.2675
No log 21.7647 370 1.4566 0.2793 1.4566 1.2069
No log 21.8824 372 1.3507 0.2506 1.3507 1.1622
No log 22.0 374 1.3616 0.1407 1.3616 1.1669
No log 22.1176 376 1.4201 0.0970 1.4201 1.1917
No log 22.2353 378 1.4705 0.0970 1.4705 1.2126
No log 22.3529 380 1.5011 0.0970 1.5011 1.2252
No log 22.4706 382 1.4913 0.1880 1.4913 1.2212
No log 22.5882 384 1.3610 0.1407 1.3610 1.1666
No log 22.7059 386 1.2859 0.1142 1.2859 1.1340
No log 22.8235 388 1.3489 0.1744 1.3489 1.1614
No log 22.9412 390 1.5244 0.2424 1.5244 1.2347
No log 23.0588 392 1.6684 0.2653 1.6684 1.2917
No log 23.1765 394 1.7069 0.2653 1.7069 1.3065
No log 23.2941 396 1.6331 0.2474 1.6331 1.2779
No log 23.4118 398 1.4977 0.1744 1.4977 1.2238
No log 23.5294 400 1.3979 0.0781 1.3979 1.1823
No log 23.6471 402 1.3591 0.1552 1.3591 1.1658
No log 23.7647 404 1.3953 0.2555 1.3953 1.1812
No log 23.8824 406 1.5681 0.2611 1.5681 1.2522
No log 24.0 408 1.6914 0.2363 1.6914 1.3005
No log 24.1176 410 1.6706 0.2906 1.6706 1.2925
No log 24.2353 412 1.5419 0.1634 1.5419 1.2417
No log 24.3529 414 1.3849 0.1552 1.3849 1.1768
No log 24.4706 416 1.3227 0.1552 1.3227 1.1501
No log 24.5882 418 1.3651 0.1552 1.3651 1.1684
No log 24.7059 420 1.5048 0.2015 1.5048 1.2267
No log 24.8235 422 1.6664 0.2062 1.6664 1.2909
No log 24.9412 424 1.7652 0.1950 1.7652 1.3286
No log 25.0588 426 1.7681 0.2062 1.7681 1.3297
No log 25.1765 428 1.7561 0.2391 1.7561 1.3252
No log 25.2941 430 1.6795 0.2653 1.6795 1.2960
No log 25.4118 432 1.5839 0.2342 1.5839 1.2585
No log 25.5294 434 1.5372 0.1814 1.5372 1.2398
No log 25.6471 436 1.5172 0.1142 1.5172 1.2317
No log 25.7647 438 1.5352 0.1486 1.5352 1.2390
No log 25.8824 440 1.5193 0.1142 1.5193 1.2326
No log 26.0 442 1.4771 0.1142 1.4771 1.2154
No log 26.1176 444 1.4575 0.1052 1.4575 1.2073
No log 26.2353 446 1.4544 0.1052 1.4544 1.2060
No log 26.3529 448 1.4995 0.0510 1.4995 1.2246
No log 26.4706 450 1.5472 0.0878 1.5472 1.2439
No log 26.5882 452 1.5856 0.1486 1.5856 1.2592
No log 26.7059 454 1.5794 0.1634 1.5794 1.2567
No log 26.8235 456 1.4910 0.1744 1.4910 1.2211
No log 26.9412 458 1.4981 0.2292 1.4981 1.2240
No log 27.0588 460 1.5214 0.2611 1.5214 1.2335
No log 27.1765 462 1.6813 0.2206 1.6813 1.2966
No log 27.2941 464 1.7428 0.2252 1.7428 1.3202
No log 27.4118 466 1.7078 0.2159 1.7078 1.3068
No log 27.5294 468 1.6739 0.1562 1.6739 1.2938
No log 27.6471 470 1.5866 0.0878 1.5866 1.2596
No log 27.7647 472 1.5118 0.1486 1.5118 1.2296
No log 27.8824 474 1.5036 0.1142 1.5036 1.2262
No log 28.0 476 1.5259 0.1486 1.5259 1.2353
No log 28.1176 478 1.4655 0.1744 1.4655 1.2106
No log 28.2353 480 1.3777 0.1142 1.3777 1.1738
No log 28.3529 482 1.3389 0.1142 1.3389 1.1571
No log 28.4706 484 1.3691 0.1142 1.3691 1.1701
No log 28.5882 486 1.4628 0.2126 1.4628 1.2095
No log 28.7059 488 1.5249 0.2184 1.5249 1.2348
No log 28.8235 490 1.5063 0.2184 1.5063 1.2273
No log 28.9412 492 1.4423 0.1228 1.4423 1.2009
No log 29.0588 494 1.4071 0.1228 1.4071 1.1862
No log 29.1765 496 1.4755 0.2474 1.4755 1.2147
No log 29.2941 498 1.4672 0.2474 1.4672 1.2113
0.267 29.4118 500 1.3991 0.1814 1.3991 1.1828
0.267 29.5294 502 1.4016 0.1744 1.4016 1.1839
0.267 29.6471 504 1.3661 0.1744 1.3661 1.1688
0.267 29.7647 506 1.3893 0.2424 1.3893 1.1787
0.267 29.8824 508 1.4721 0.2424 1.4721 1.2133
0.267 30.0 510 1.4838 0.2424 1.4838 1.2181
0.267 30.1176 512 1.6102 0.3117 1.6102 1.2689
0.267 30.2353 514 1.6458 0.3117 1.6458 1.2829
0.267 30.3529 516 1.5810 0.2752 1.5810 1.2574
0.267 30.4706 518 1.4660 0.1142 1.4660 1.2108
0.267 30.5882 520 1.3825 0.0401 1.3825 1.1758
0.267 30.7059 522 1.3373 0.0401 1.3373 1.1564
0.267 30.8235 524 1.2777 0.0401 1.2777 1.1304
0.267 30.9412 526 1.3044 0.0401 1.3044 1.1421

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k7_task5_organization

Finetuned
(4019)
this model