ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k5_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2819
  • Qwk: 0.1345
  • Mse: 1.2819
  • Rmse: 1.1322

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.08 2 4.6431 -0.0020 4.6431 2.1548
No log 0.16 4 2.7124 0.0025 2.7124 1.6469
No log 0.24 6 1.9882 0.0169 1.9882 1.4100
No log 0.32 8 1.3786 0.0398 1.3786 1.1741
No log 0.4 10 1.4618 0.0872 1.4618 1.2091
No log 0.48 12 1.5905 0.0635 1.5905 1.2612
No log 0.56 14 2.2608 -0.0476 2.2608 1.5036
No log 0.64 16 2.2645 -0.0532 2.2645 1.5048
No log 0.72 18 1.6082 -0.0448 1.6082 1.2682
No log 0.8 20 1.2701 0.0613 1.2701 1.1270
No log 0.88 22 1.2837 0.1142 1.2837 1.1330
No log 0.96 24 1.3169 0.1271 1.3169 1.1476
No log 1.04 26 1.3290 0.1470 1.3290 1.1528
No log 1.12 28 1.3478 0.1109 1.3478 1.1609
No log 1.2 30 1.3585 -0.0401 1.3585 1.1656
No log 1.28 32 1.4170 -0.0452 1.4170 1.1904
No log 1.3600 34 1.4918 0.0462 1.4918 1.2214
No log 1.44 36 1.4962 0.0992 1.4962 1.2232
No log 1.52 38 1.5320 0.0732 1.5320 1.2377
No log 1.6 40 1.5059 0.1282 1.5059 1.2271
No log 1.6800 42 1.4063 0.1282 1.4063 1.1859
No log 1.76 44 1.3327 0.1346 1.3327 1.1544
No log 1.8400 46 1.2896 0.0768 1.2896 1.1356
No log 1.92 48 1.1978 0.2200 1.1978 1.0945
No log 2.0 50 1.1423 0.1935 1.1423 1.0688
No log 2.08 52 1.1961 0.2223 1.1961 1.0937
No log 2.16 54 1.2965 0.1848 1.2965 1.1386
No log 2.24 56 1.4154 0.0497 1.4154 1.1897
No log 2.32 58 1.5288 -0.0096 1.5288 1.2364
No log 2.4 60 1.5691 -0.0010 1.5691 1.2526
No log 2.48 62 1.5216 0.0497 1.5216 1.2335
No log 2.56 64 1.3310 0.1108 1.3310 1.1537
No log 2.64 66 1.3040 0.0694 1.3040 1.1419
No log 2.7200 68 1.2748 0.1650 1.2748 1.1291
No log 2.8 70 1.2782 0.1338 1.2782 1.1306
No log 2.88 72 1.4682 0.1894 1.4682 1.2117
No log 2.96 74 1.3624 0.0892 1.3624 1.1672
No log 3.04 76 1.2111 0.2640 1.2111 1.1005
No log 3.12 78 1.2355 0.1706 1.2355 1.1115
No log 3.2 80 1.2182 0.1935 1.2182 1.1037
No log 3.2800 82 1.2632 0.1614 1.2632 1.1239
No log 3.36 84 1.6892 0.1331 1.6892 1.2997
No log 3.44 86 1.8878 0.1163 1.8878 1.3740
No log 3.52 88 1.6784 0.1281 1.6784 1.2955
No log 3.6 90 1.3920 0.1046 1.3920 1.1798
No log 3.68 92 1.3776 0.1109 1.3776 1.1737
No log 3.76 94 1.4007 0.1500 1.4007 1.1835
No log 3.84 96 1.3125 0.0811 1.3125 1.1457
No log 3.92 98 1.3422 0.1448 1.3422 1.1585
No log 4.0 100 1.4449 0.2006 1.4449 1.2021
No log 4.08 102 1.3819 0.1708 1.3819 1.1755
No log 4.16 104 1.2085 0.1538 1.2085 1.0993
No log 4.24 106 1.1679 0.1114 1.1679 1.0807
No log 4.32 108 1.2022 0.1217 1.2022 1.0964
No log 4.4 110 1.2178 0.1217 1.2178 1.1035
No log 4.48 112 1.3509 0.1313 1.3509 1.1623
No log 4.5600 114 1.5724 0.2424 1.5724 1.2540
No log 4.64 116 1.5514 0.2632 1.5514 1.2455
No log 4.72 118 1.4086 0.2239 1.4086 1.1868
No log 4.8 120 1.4175 0.1708 1.4175 1.1906
No log 4.88 122 1.5670 0.2873 1.5670 1.2518
No log 4.96 124 1.5035 0.2275 1.5035 1.2262
No log 5.04 126 1.3407 0.1462 1.3407 1.1579
No log 5.12 128 1.4000 0.1159 1.4000 1.1832
No log 5.2 130 1.5637 0.3026 1.5637 1.2505
No log 5.28 132 1.5106 0.1698 1.5106 1.2291
No log 5.36 134 1.4923 0.1608 1.4923 1.2216
No log 5.44 136 1.6161 0.1667 1.6161 1.2712
No log 5.52 138 1.5283 0.1196 1.5283 1.2362
No log 5.6 140 1.3209 0.1020 1.3209 1.1493
No log 5.68 142 1.2909 0.2188 1.2909 1.1362
No log 5.76 144 1.3277 0.1183 1.3277 1.1522
No log 5.84 146 1.5509 0.2020 1.5509 1.2454
No log 5.92 148 1.7111 0.1657 1.7111 1.3081
No log 6.0 150 1.5912 0.2375 1.5912 1.2614
No log 6.08 152 1.3422 0.1500 1.3422 1.1585
No log 6.16 154 1.2960 0.1442 1.2960 1.1384
No log 6.24 156 1.3524 0.2132 1.3524 1.1629
No log 6.32 158 1.3949 0.2250 1.3949 1.1811
No log 6.4 160 1.5512 0.2206 1.5512 1.2455
No log 6.48 162 1.9320 0.1785 1.9320 1.3900
No log 6.5600 164 1.9933 0.1785 1.9933 1.4119
No log 6.64 166 1.7955 0.2213 1.7955 1.3399
No log 6.72 168 1.4300 0.2589 1.4300 1.1958
No log 6.8 170 1.3439 0.1795 1.3439 1.1593
No log 6.88 172 1.3103 0.2009 1.3103 1.1447
No log 6.96 174 1.4899 0.2060 1.4899 1.2206
No log 7.04 176 1.6256 0.2465 1.6256 1.2750
No log 7.12 178 1.6809 0.2708 1.6809 1.2965
No log 7.2 180 1.5422 0.2406 1.5422 1.2419
No log 7.28 182 1.4577 0.3106 1.4577 1.2074
No log 7.36 184 1.6890 0.2540 1.6890 1.2996
No log 7.44 186 1.8066 0.1946 1.8066 1.3441
No log 7.52 188 1.5415 0.2348 1.5415 1.2416
No log 7.6 190 1.3681 0.3025 1.3681 1.1697
No log 7.68 192 1.4037 0.2674 1.4037 1.1848
No log 7.76 194 1.4914 0.2206 1.4914 1.2212
No log 7.84 196 1.3543 0.2074 1.3543 1.1637
No log 7.92 198 1.4142 0.1763 1.4142 1.1892
No log 8.0 200 1.3370 0.2602 1.3370 1.1563
No log 8.08 202 1.2792 0.1750 1.2792 1.1310
No log 8.16 204 1.2418 0.1379 1.2418 1.1144
No log 8.24 206 1.2288 0.1482 1.2288 1.1085
No log 8.32 208 1.3351 0.0541 1.3351 1.1555
No log 8.4 210 1.4359 0.0887 1.4359 1.1983
No log 8.48 212 1.4676 0.0887 1.4676 1.2114
No log 8.56 214 1.5145 0.0273 1.5145 1.2306
No log 8.64 216 1.5406 0.0851 1.5406 1.2412
No log 8.72 218 1.4499 0.1980 1.4499 1.2041
No log 8.8 220 1.3809 0.2263 1.3809 1.1751
No log 8.88 222 1.5549 0.1386 1.5549 1.2470
No log 8.96 224 1.6245 0.1884 1.6245 1.2745
No log 9.04 226 1.5527 0.2009 1.5527 1.2461
No log 9.12 228 1.2551 0.1707 1.2551 1.1203
No log 9.2 230 1.1832 0.1761 1.1832 1.0877
No log 9.28 232 1.2979 0.2184 1.2979 1.1393
No log 9.36 234 1.6031 0.2079 1.6031 1.2661
No log 9.44 236 1.8312 0.1282 1.8312 1.3532
No log 9.52 238 1.9665 0.1401 1.9665 1.4023
No log 9.6 240 1.7777 0.1282 1.7777 1.3333
No log 9.68 242 1.7101 0.1555 1.7101 1.3077
No log 9.76 244 1.5510 0.2574 1.5510 1.2454
No log 9.84 246 1.6702 0.1135 1.6702 1.2924
No log 9.92 248 1.8438 0.1634 1.8438 1.3579
No log 10.0 250 2.0678 0.0653 2.0678 1.4380
No log 10.08 252 2.0633 0.0653 2.0633 1.4364
No log 10.16 254 1.7572 0.1135 1.7572 1.3256
No log 10.24 256 1.5440 0.1283 1.5440 1.2426
No log 10.32 258 1.6568 0.1009 1.6568 1.2872
No log 10.4 260 1.6817 0.1232 1.6817 1.2968
No log 10.48 262 1.6303 0.1713 1.6303 1.2768
No log 10.56 264 1.6732 0.1436 1.6732 1.2935
No log 10.64 266 1.8116 0.0959 1.8116 1.3460
No log 10.72 268 1.8948 0.1197 1.8948 1.3765
No log 10.8 270 1.8172 0.1555 1.8172 1.3480
No log 10.88 272 1.5712 0.2009 1.5712 1.2535
No log 10.96 274 1.4680 0.1979 1.4680 1.2116
No log 11.04 276 1.5099 0.2574 1.5099 1.2288
No log 11.12 278 1.5994 0.2047 1.5994 1.2647
No log 11.2 280 1.7558 0.1736 1.7558 1.3250
No log 11.28 282 1.7703 0.1953 1.7703 1.3305
No log 11.36 284 1.7134 0.1616 1.7134 1.3090
No log 11.44 286 1.6048 0.1884 1.6048 1.2668
No log 11.52 288 1.3776 0.2905 1.3776 1.1737
No log 11.6 290 1.2981 0.1217 1.2981 1.1393
No log 11.68 292 1.3249 0.2097 1.3249 1.1510
No log 11.76 294 1.4140 0.2754 1.4140 1.1891
No log 11.84 296 1.4651 0.2105 1.4651 1.2104
No log 11.92 298 1.4490 0.2289 1.4490 1.2037
No log 12.0 300 1.3783 0.2647 1.3783 1.1740
No log 12.08 302 1.3628 0.2647 1.3628 1.1674
No log 12.16 304 1.3985 0.3152 1.3985 1.1826
No log 12.24 306 1.4559 0.2455 1.4559 1.2066
No log 12.32 308 1.4108 0.2544 1.4108 1.1878
No log 12.4 310 1.2771 0.2570 1.2771 1.1301
No log 12.48 312 1.2792 0.2570 1.2792 1.1310
No log 12.56 314 1.2985 0.1596 1.2985 1.1395
No log 12.64 316 1.3510 0.1703 1.3510 1.1623
No log 12.72 318 1.5371 0.2144 1.5371 1.2398
No log 12.8 320 1.8289 0.1697 1.8289 1.3524
No log 12.88 322 1.8328 0.2066 1.8328 1.3538
No log 12.96 324 1.5998 0.2217 1.5998 1.2648
No log 13.04 326 1.3232 0.1500 1.3232 1.1503
No log 13.12 328 1.2365 0.1641 1.2365 1.1120
No log 13.2 330 1.3315 0.1557 1.3315 1.1539
No log 13.28 332 1.6501 0.2317 1.6501 1.2846
No log 13.36 334 2.0104 0.1476 2.0104 1.4179
No log 13.44 336 1.9713 0.1304 1.9713 1.4040
No log 13.52 338 1.7649 0.1923 1.7649 1.3285
No log 13.6 340 1.5033 -0.0072 1.5033 1.2261
No log 13.68 342 1.3387 0.0541 1.3387 1.1570
No log 13.76 344 1.3017 0.0442 1.3017 1.1409
No log 13.84 346 1.3506 0.0442 1.3506 1.1622
No log 13.92 348 1.5556 0.2359 1.5556 1.2472
No log 14.0 350 1.6920 0.2285 1.6920 1.3008
No log 14.08 352 1.6907 0.2251 1.6907 1.3003
No log 14.16 354 1.5726 0.2009 1.5726 1.2540
No log 14.24 356 1.4186 0.1592 1.4186 1.1911
No log 14.32 358 1.3397 0.0887 1.3397 1.1575
No log 14.4 360 1.3425 0.0702 1.3425 1.1587
No log 14.48 362 1.4351 0.2410 1.4351 1.1980
No log 14.56 364 1.6287 0.2009 1.6287 1.2762
No log 14.64 366 1.7976 0.1767 1.7976 1.3407
No log 14.72 368 1.7910 0.1767 1.7910 1.3383
No log 14.8 370 1.7132 0.1884 1.7132 1.3089
No log 14.88 372 1.5272 0.2359 1.5272 1.2358
No log 14.96 374 1.3778 0.2213 1.3778 1.1738
No log 15.04 376 1.2871 0.1122 1.2871 1.1345
No log 15.12 378 1.2709 0.0789 1.2709 1.1273
No log 15.2 380 1.3090 0.1498 1.3090 1.1441
No log 15.28 382 1.4098 0.2359 1.4098 1.1873
No log 15.36 384 1.5940 0.2217 1.5940 1.2625
No log 15.44 386 1.6558 0.2285 1.6558 1.2868
No log 15.52 388 1.5662 0.2217 1.5662 1.2515
No log 15.6 390 1.4439 0.2359 1.4439 1.2016
No log 15.68 392 1.4176 0.2065 1.4176 1.1906
No log 15.76 394 1.4618 0.1929 1.4618 1.2091
No log 15.84 396 1.4635 0.2217 1.4635 1.2098
No log 15.92 398 1.4604 0.2392 1.4604 1.2085
No log 16.0 400 1.3737 0.3238 1.3737 1.1720
No log 16.08 402 1.3765 0.3238 1.3765 1.1733
No log 16.16 404 1.3468 0.3238 1.3468 1.1605
No log 16.24 406 1.2857 0.2707 1.2857 1.1339
No log 16.32 408 1.3107 0.2707 1.3107 1.1449
No log 16.4 410 1.4689 0.2217 1.4689 1.2120
No log 16.48 412 1.5418 0.2217 1.5418 1.2417
No log 16.56 414 1.4941 0.2105 1.4941 1.2223
No log 16.64 416 1.3463 0.0702 1.3463 1.1603
No log 16.72 418 1.3067 0.0541 1.3067 1.1431
No log 16.8 420 1.2663 0.0904 1.2663 1.1253
No log 16.88 422 1.2973 0.0454 1.2973 1.1390
No log 16.96 424 1.3686 0.0887 1.3686 1.1699
No log 17.04 426 1.4657 0.1698 1.4657 1.2107
No log 17.12 428 1.5110 0.2763 1.5110 1.2292
No log 17.2 430 1.4295 0.1283 1.4295 1.1956
No log 17.28 432 1.3590 0.0702 1.3590 1.1658
No log 17.36 434 1.3960 0.0702 1.3960 1.1815
No log 17.44 436 1.4929 0.2512 1.4929 1.2218
No log 17.52 438 1.5394 0.2217 1.5394 1.2407
No log 17.6 440 1.5811 0.2217 1.5811 1.2574
No log 17.68 442 1.5183 0.2217 1.5183 1.2322
No log 17.76 444 1.5533 0.2217 1.5533 1.2463
No log 17.84 446 1.5224 0.2217 1.5224 1.2338
No log 17.92 448 1.4269 0.2604 1.4269 1.1945
No log 18.0 450 1.3145 0.2173 1.3145 1.1465
No log 18.08 452 1.2219 0.1345 1.2219 1.1054
No log 18.16 454 1.2264 0.1889 1.2264 1.1074
No log 18.24 456 1.3122 0.2935 1.3122 1.1455
No log 18.32 458 1.4300 0.2392 1.4300 1.1958
No log 18.4 460 1.5395 0.2084 1.5395 1.2407
No log 18.48 462 1.5455 0.2120 1.5455 1.2432
No log 18.56 464 1.4418 0.2392 1.4418 1.2008
No log 18.64 466 1.3322 0.2544 1.3322 1.1542
No log 18.72 468 1.3700 0.2410 1.3700 1.1705
No log 18.8 470 1.4831 0.1713 1.4831 1.2178
No log 18.88 472 1.5344 0.1228 1.5344 1.2387
No log 18.96 474 1.4464 0.1935 1.4464 1.2026
No log 19.04 476 1.2717 0.0541 1.2717 1.1277
No log 19.12 478 1.1740 0.1247 1.1740 1.0835
No log 19.2 480 1.1638 0.1482 1.1638 1.0788
No log 19.28 482 1.1974 0.1076 1.1974 1.0943
No log 19.36 484 1.3000 0.1219 1.3000 1.1402
No log 19.44 486 1.4760 0.1970 1.4760 1.2149
No log 19.52 488 1.5112 0.2217 1.5112 1.2293
No log 19.6 490 1.4411 0.2789 1.4411 1.2005
No log 19.68 492 1.3094 0.2173 1.3094 1.1443
No log 19.76 494 1.2120 0.25 1.2120 1.1009
No log 19.84 496 1.2306 0.2809 1.2306 1.1093
No log 19.92 498 1.3223 0.3202 1.3223 1.1499
0.3569 20.0 500 1.5294 0.1970 1.5294 1.2367
0.3569 20.08 502 1.6503 0.1802 1.6503 1.2846
0.3569 20.16 504 1.6817 0.1547 1.6817 1.2968
0.3569 20.24 506 1.5753 0.1162 1.5753 1.2551
0.3569 20.32 508 1.3905 0.1031 1.3905 1.1792
0.3569 20.4 510 1.2819 0.1345 1.2819 1.1322

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k5_task2_organization

Finetuned
(4019)
this model