ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k16_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.5139
  • Qwk: 0.0401
  • Mse: 1.5139
  • Rmse: 1.2304

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 3.9297 -0.0091 3.9297 1.9823
No log 0.1053 4 2.2644 0.0093 2.2644 1.5048
No log 0.1579 6 1.4707 -0.0342 1.4707 1.2127
No log 0.2105 8 1.1646 0.1901 1.1646 1.0792
No log 0.2632 10 1.1048 0.1891 1.1048 1.0511
No log 0.3158 12 1.4166 -0.0720 1.4166 1.1902
No log 0.3684 14 1.5278 -0.0657 1.5278 1.2361
No log 0.4211 16 1.1902 0.2268 1.1902 1.0910
No log 0.4737 18 1.0738 0.2492 1.0738 1.0362
No log 0.5263 20 1.1138 0.1107 1.1138 1.0554
No log 0.5789 22 1.0448 0.2061 1.0448 1.0222
No log 0.6316 24 1.0492 0.3082 1.0492 1.0243
No log 0.6842 26 1.0534 0.2834 1.0534 1.0263
No log 0.7368 28 0.9923 0.3457 0.9923 0.9961
No log 0.7895 30 1.1923 0.0884 1.1923 1.0919
No log 0.8421 32 1.5746 -0.2749 1.5746 1.2548
No log 0.8947 34 1.4644 -0.1478 1.4644 1.2101
No log 0.9474 36 1.2039 0.0293 1.2039 1.0972
No log 1.0 38 1.1307 0.0855 1.1307 1.0633
No log 1.0526 40 1.3227 -0.0057 1.3227 1.1501
No log 1.1053 42 1.6379 -0.0797 1.6379 1.2798
No log 1.1579 44 1.6627 -0.0210 1.6627 1.2895
No log 1.2105 46 1.8100 -0.0904 1.8100 1.3454
No log 1.2632 48 1.8027 -0.0274 1.8027 1.3427
No log 1.3158 50 1.8947 -0.0763 1.8947 1.3765
No log 1.3684 52 1.8409 0.0178 1.8409 1.3568
No log 1.4211 54 1.6384 0.1545 1.6384 1.2800
No log 1.4737 56 1.7826 -0.1017 1.7826 1.3351
No log 1.5263 58 1.7611 -0.1364 1.7611 1.3271
No log 1.5789 60 1.6069 0.0791 1.6069 1.2676
No log 1.6316 62 1.6913 -0.0381 1.6913 1.3005
No log 1.6842 64 1.6415 -0.0192 1.6415 1.2812
No log 1.7368 66 1.4119 0.1462 1.4119 1.1882
No log 1.7895 68 1.3219 0.1952 1.3219 1.1497
No log 1.8421 70 1.4256 0.1462 1.4256 1.1940
No log 1.8947 72 1.5222 0.1298 1.5222 1.2338
No log 1.9474 74 1.4131 0.1538 1.4131 1.1887
No log 2.0 76 1.2267 0.0464 1.2267 1.1076
No log 2.0526 78 1.0881 0.0496 1.0881 1.0431
No log 2.1053 80 1.0730 0.0618 1.0730 1.0358
No log 2.1579 82 1.1872 0.0527 1.1872 1.0896
No log 2.2105 84 1.5001 -0.0468 1.5001 1.2248
No log 2.2632 86 1.7777 -0.1153 1.7777 1.3333
No log 2.3158 88 1.7310 -0.0084 1.7310 1.3157
No log 2.3684 90 1.7927 0.1211 1.7927 1.3389
No log 2.4211 92 1.8897 0.1559 1.8897 1.3747
No log 2.4737 94 1.7848 0.1918 1.7848 1.3360
No log 2.5263 96 1.6565 0.1752 1.6565 1.2871
No log 2.5789 98 1.4749 0.2184 1.4749 1.2145
No log 2.6316 100 1.6391 0.1950 1.6391 1.2803
No log 2.6842 102 1.8270 0.1911 1.8270 1.3517
No log 2.7368 104 1.8310 0.1607 1.8310 1.3532
No log 2.7895 106 1.7616 0.1423 1.7616 1.3273
No log 2.8421 108 1.5840 0.2372 1.5840 1.2586
No log 2.8947 110 1.6116 0.1769 1.6116 1.2695
No log 2.9474 112 1.7428 0.2005 1.7428 1.3201
No log 3.0 114 1.5862 0.1832 1.5862 1.2594
No log 3.0526 116 1.4767 0.2184 1.4767 1.2152
No log 3.1053 118 1.4914 0.2474 1.4914 1.2212
No log 3.1579 120 1.4850 0.2611 1.4850 1.2186
No log 3.2105 122 1.6911 0.2527 1.6911 1.3004
No log 3.2632 124 1.7546 0.2296 1.7546 1.3246
No log 3.3158 126 1.5766 0.2522 1.5766 1.2556
No log 3.3684 128 1.3352 0.1744 1.3352 1.1555
No log 3.4211 130 1.2309 0.1142 1.2309 1.1095
No log 3.4737 132 1.2192 0.1142 1.2192 1.1042
No log 3.5263 134 1.2829 0.1407 1.2829 1.1327
No log 3.5789 136 1.3882 0.1814 1.3882 1.1782
No log 3.6316 138 1.4315 0.1814 1.4315 1.1965
No log 3.6842 140 1.3254 0.1744 1.3254 1.1513
No log 3.7368 142 1.2270 0.1538 1.2270 1.1077
No log 3.7895 144 1.2628 0.1725 1.2628 1.1237
No log 3.8421 146 1.4160 0.1744 1.4160 1.1899
No log 3.8947 148 1.6286 0.1298 1.6286 1.2762
No log 3.9474 150 1.7016 0.0500 1.7016 1.3045
No log 4.0 152 1.6057 0.0694 1.6057 1.2672
No log 4.0526 154 1.4151 0.1228 1.4151 1.1896
No log 4.1053 156 1.1929 0.1142 1.1929 1.0922
No log 4.1579 158 1.0784 0.1886 1.0784 1.0385
No log 4.2105 160 1.0832 0.2506 1.0832 1.0408
No log 4.2632 162 1.1442 0.2372 1.1442 1.0697
No log 4.3158 164 1.2599 0.2372 1.2599 1.1224
No log 4.3684 166 1.4571 0.2752 1.4571 1.2071
No log 4.4211 168 1.6631 0.2252 1.6631 1.2896
No log 4.4737 170 1.6241 0.2363 1.6241 1.2744
No log 4.5263 172 1.4599 0.2126 1.4599 1.2083
No log 4.5789 174 1.3276 0.2126 1.3276 1.1522
No log 4.6316 176 1.2589 0.2372 1.2589 1.1220
No log 4.6842 178 1.2213 0.2372 1.2213 1.1051
No log 4.7368 180 1.2310 0.1744 1.2310 1.1095
No log 4.7895 182 1.2728 0.1744 1.2728 1.1282
No log 4.8421 184 1.2222 0.1142 1.2222 1.1055
No log 4.8947 186 1.1800 0.1407 1.1800 1.0863
No log 4.9474 188 1.2156 0.1744 1.2156 1.1026
No log 5.0 190 1.2664 0.1744 1.2664 1.1254
No log 5.0526 192 1.3711 0.2372 1.3711 1.1709
No log 5.1053 194 1.5439 0.2568 1.5439 1.2425
No log 5.1579 196 1.7491 0.2840 1.7491 1.3225
No log 5.2105 198 1.9123 0.2222 1.9123 1.3829
No log 5.2632 200 2.0071 0.2172 2.0071 1.4167
No log 5.3158 202 1.9567 0.1925 1.9567 1.3988
No log 5.3684 204 1.8003 0.2007 1.8003 1.3418
No log 5.4211 206 1.6237 0.2525 1.6237 1.2743
No log 5.4737 208 1.3243 0.2424 1.3243 1.1508
No log 5.5263 210 1.2120 0.1744 1.2120 1.1009
No log 5.5789 212 1.2682 0.1744 1.2682 1.1262
No log 5.6316 214 1.3827 0.1744 1.3827 1.1759
No log 5.6842 216 1.5327 0.2126 1.5327 1.2380
No log 5.7368 218 1.6381 0.2117 1.6381 1.2799
No log 5.7895 220 1.5877 0.2062 1.5877 1.2600
No log 5.8421 222 1.5552 0.2062 1.5552 1.2471
No log 5.8947 224 1.5181 0.2522 1.5181 1.2321
No log 5.9474 226 1.3977 0.2126 1.3977 1.1823
No log 6.0 228 1.4579 0.2184 1.4579 1.2074
No log 6.0526 230 1.6840 0.2270 1.6840 1.2977
No log 6.1053 232 1.7565 0.3095 1.7565 1.3253
No log 6.1579 234 1.6280 0.2806 1.6280 1.2759
No log 6.2105 236 1.4439 0.2184 1.4439 1.2016
No log 6.2632 238 1.3651 0.2126 1.3651 1.1684
No log 6.3158 240 1.3989 0.2126 1.3989 1.1827
No log 6.3684 242 1.5432 0.2611 1.5432 1.2423
No log 6.4211 244 1.5809 0.2906 1.5809 1.2573
No log 6.4737 246 1.5051 0.2793 1.5051 1.2268
No log 6.5263 248 1.4096 0.2372 1.4096 1.1873
No log 6.5789 250 1.3961 0.1744 1.3961 1.1816
No log 6.6316 252 1.3757 0.1814 1.3757 1.1729
No log 6.6842 254 1.3679 0.2126 1.3679 1.1696
No log 6.7368 256 1.3306 0.1744 1.3306 1.1535
No log 6.7895 258 1.2798 0.1744 1.2798 1.1313
No log 6.8421 260 1.2702 0.1744 1.2702 1.1270
No log 6.8947 262 1.3957 0.1744 1.3957 1.1814
No log 6.9474 264 1.5052 0.2372 1.5052 1.2268
No log 7.0 266 1.5316 0.2065 1.5316 1.2376
No log 7.0526 268 1.4798 0.1744 1.4798 1.2165
No log 7.1053 270 1.3768 0.0781 1.3768 1.1734
No log 7.1579 272 1.3837 0.0781 1.3837 1.1763
No log 7.2105 274 1.4395 0.1407 1.4395 1.1998
No log 7.2632 276 1.5374 0.2522 1.5374 1.2399
No log 7.3158 278 1.5189 0.2611 1.5189 1.2324
No log 7.3684 280 1.4733 0.2342 1.4733 1.2138
No log 7.4211 282 1.5213 0.2653 1.5213 1.2334
No log 7.4737 284 1.5974 0.2566 1.5974 1.2639
No log 7.5263 286 1.5360 0.2653 1.5360 1.2394
No log 7.5789 288 1.3785 0.2474 1.3785 1.1741
No log 7.6316 290 1.2328 0.1407 1.2328 1.1103
No log 7.6842 292 1.2197 0.1407 1.2197 1.1044
No log 7.7368 294 1.2750 0.1407 1.2750 1.1292
No log 7.7895 296 1.3850 0.2424 1.3850 1.1769
No log 7.8421 298 1.4062 0.2424 1.4062 1.1858
No log 7.8947 300 1.3927 0.1407 1.3927 1.1801
No log 7.9474 302 1.4390 0.1407 1.4390 1.1996
No log 8.0 304 1.5794 0.2653 1.5794 1.2568
No log 8.0526 306 1.8038 0.2206 1.8038 1.3431
No log 8.1053 308 2.0104 0.0915 2.0104 1.4179
No log 8.1579 310 2.0191 0.0915 2.0191 1.4209
No log 8.2105 312 1.9120 0.1313 1.9120 1.3828
No log 8.2632 314 1.6732 0.2482 1.6732 1.2935
No log 8.3158 316 1.3978 0.2126 1.3978 1.1823
No log 8.3684 318 1.2624 0.0781 1.2624 1.1235
No log 8.4211 320 1.2196 0.0401 1.2196 1.1043
No log 8.4737 322 1.2527 0.0401 1.2527 1.1192
No log 8.5263 324 1.3804 0.0781 1.3804 1.1749
No log 8.5789 326 1.5970 0.2184 1.5970 1.2637
No log 8.6316 328 1.7462 0.2606 1.7462 1.3214
No log 8.6842 330 1.8395 0.2236 1.8395 1.3563
No log 8.7368 332 1.9571 0.2317 1.9571 1.3990
No log 8.7895 334 1.9411 0.2317 1.9411 1.3932
No log 8.8421 336 1.8324 0.2566 1.8324 1.3537
No log 8.8947 338 1.6554 0.2653 1.6554 1.2866
No log 8.9474 340 1.5323 0.2065 1.5323 1.2379
No log 9.0 342 1.4333 0.1744 1.4333 1.1972
No log 9.0526 344 1.4335 0.1744 1.4335 1.1973
No log 9.1053 346 1.4070 0.0781 1.4070 1.1862
No log 9.1579 348 1.4165 0.0781 1.4165 1.1902
No log 9.2105 350 1.5179 0.2474 1.5179 1.2320
No log 9.2632 352 1.5659 0.2474 1.5659 1.2514
No log 9.3158 354 1.5811 0.2482 1.5811 1.2574
No log 9.3684 356 1.4755 0.2437 1.4755 1.2147
No log 9.4211 358 1.5155 0.2437 1.5155 1.2310
No log 9.4737 360 1.6228 0.2437 1.6228 1.2739
No log 9.5263 362 1.6622 0.2653 1.6622 1.2893
No log 9.5789 364 1.5729 0.2239 1.5729 1.2541
No log 9.6316 366 1.5349 0.2372 1.5349 1.2389
No log 9.6842 368 1.5920 0.2292 1.5920 1.2617
No log 9.7368 370 1.5865 0.2292 1.5865 1.2596
No log 9.7895 372 1.4858 0.2424 1.4858 1.2189
No log 9.8421 374 1.3642 0.1744 1.3642 1.1680
No log 9.8947 376 1.2817 0.0401 1.2817 1.1321
No log 9.9474 378 1.3018 0.0401 1.3018 1.1409
No log 10.0 380 1.4051 0.1744 1.4051 1.1854
No log 10.0526 382 1.5001 0.2474 1.5001 1.2248
No log 10.1053 384 1.5417 0.2832 1.5417 1.2416
No log 10.1579 386 1.4758 0.2474 1.4758 1.2148
No log 10.2105 388 1.4637 0.2522 1.4637 1.2098
No log 10.2632 390 1.4660 0.2568 1.4660 1.2108
No log 10.3158 392 1.5715 0.2733 1.5715 1.2536
No log 10.3684 394 1.6138 0.2566 1.6138 1.2704
No log 10.4211 396 1.7155 0.2644 1.7155 1.3098
No log 10.4737 398 1.7725 0.2296 1.7725 1.3314
No log 10.5263 400 1.6584 0.2342 1.6584 1.2878
No log 10.5789 402 1.4174 0.1744 1.4174 1.1905
No log 10.6316 404 1.2911 0.0401 1.2911 1.1363
No log 10.6842 406 1.2856 0.0401 1.2856 1.1339
No log 10.7368 408 1.3198 0.0401 1.3198 1.1488
No log 10.7895 410 1.3973 0.0781 1.3973 1.1821
No log 10.8421 412 1.5352 0.2372 1.5352 1.2390
No log 10.8947 414 1.6063 0.2239 1.6063 1.2674
No log 10.9474 416 1.5863 0.2184 1.5863 1.2595
No log 11.0 418 1.5779 0.2474 1.5779 1.2561
No log 11.0526 420 1.5380 0.2184 1.5380 1.2402
No log 11.1053 422 1.4525 0.1744 1.4525 1.2052
No log 11.1579 424 1.3534 0.1407 1.3534 1.1634
No log 11.2105 426 1.3738 0.1744 1.3738 1.1721
No log 11.2632 428 1.5193 0.2869 1.5193 1.2326
No log 11.3158 430 1.6551 0.2906 1.6551 1.2865
No log 11.3684 432 1.6602 0.2869 1.6602 1.2885
No log 11.4211 434 1.5091 0.2832 1.5091 1.2284
No log 11.4737 436 1.4459 0.2568 1.4459 1.2025
No log 11.5263 438 1.3842 0.2065 1.3842 1.1765
No log 11.5789 440 1.3844 0.1744 1.3844 1.1766
No log 11.6316 442 1.3852 0.1142 1.3852 1.1770
No log 11.6842 444 1.4396 0.1142 1.4396 1.1998
No log 11.7368 446 1.4940 0.1744 1.4940 1.2223
No log 11.7895 448 1.5478 0.2372 1.5478 1.2441
No log 11.8421 450 1.6907 0.2437 1.6907 1.3003
No log 11.8947 452 1.7586 0.2317 1.7586 1.3261
No log 11.9474 454 1.7151 0.2611 1.7151 1.3096
No log 12.0 456 1.5614 0.1744 1.5614 1.2495
No log 12.0526 458 1.4701 0.1744 1.4701 1.2125
No log 12.1053 460 1.4480 0.2065 1.4480 1.2033
No log 12.1579 462 1.4346 0.2065 1.4346 1.1977
No log 12.2105 464 1.4134 0.2455 1.4134 1.1889
No log 12.2632 466 1.4295 0.2455 1.4295 1.1956
No log 12.3158 468 1.4316 0.2065 1.4316 1.1965
No log 12.3684 470 1.5124 0.1943 1.5124 1.2298
No log 12.4211 472 1.5521 0.2568 1.5521 1.2458
No log 12.4737 474 1.5410 0.2568 1.5410 1.2414
No log 12.5263 476 1.4805 0.2568 1.4805 1.2168
No log 12.5789 478 1.4101 0.1880 1.4101 1.1875
No log 12.6316 480 1.4518 0.1880 1.4518 1.2049
No log 12.6842 482 1.4576 0.1880 1.4576 1.2073
No log 12.7368 484 1.4095 0.1744 1.4095 1.1872
No log 12.7895 486 1.3439 0.1407 1.3439 1.1592
No log 12.8421 488 1.3151 0.0401 1.3151 1.1468
No log 12.8947 490 1.3256 0.1744 1.3256 1.1513
No log 12.9474 492 1.3907 0.1744 1.3907 1.1793
No log 13.0 494 1.4417 0.2065 1.4417 1.2007
No log 13.0526 496 1.4405 0.1744 1.4405 1.2002
No log 13.1053 498 1.4298 0.1744 1.4298 1.1958
0.246 13.1579 500 1.4563 0.1744 1.4563 1.2068
0.246 13.2105 502 1.5172 0.1744 1.5172 1.2318
0.246 13.2632 504 1.5144 0.1407 1.5144 1.2306
0.246 13.3158 506 1.5173 0.0401 1.5173 1.2318
0.246 13.3684 508 1.5028 0.0401 1.5028 1.2259
0.246 13.4211 510 1.5139 0.0401 1.5139 1.2304

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k16_task5_organization

Finetuned
(4019)
this model