ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k3_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1380
  • Qwk: 0.4516
  • Mse: 1.1380
  • Rmse: 1.0668

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 6.9406 0.0242 6.9406 2.6345
No log 0.3077 4 4.8996 0.0738 4.8996 2.2135
No log 0.4615 6 4.1465 -0.0098 4.1465 2.0363
No log 0.6154 8 4.4566 -0.0457 4.4566 2.1111
No log 0.7692 10 3.2699 -0.0226 3.2699 1.8083
No log 0.9231 12 2.1233 0.2734 2.1233 1.4571
No log 1.0769 14 1.7910 0.3415 1.7910 1.3383
No log 1.2308 16 1.6382 0.2857 1.6382 1.2799
No log 1.3846 18 1.6285 0.2593 1.6285 1.2761
No log 1.5385 20 1.6084 0.1887 1.6084 1.2682
No log 1.6923 22 1.6443 0.2243 1.6443 1.2823
No log 1.8462 24 1.8600 0.2203 1.8600 1.3638
No log 2.0 26 2.2226 0.1583 2.2226 1.4908
No log 2.1538 28 2.6199 0.0519 2.6199 1.6186
No log 2.3077 30 2.4185 0.1060 2.4185 1.5552
No log 2.4615 32 2.0211 0.2667 2.0211 1.4217
No log 2.6154 34 1.7913 0.3770 1.7913 1.3384
No log 2.7692 36 1.6899 0.3697 1.6899 1.3000
No log 2.9231 38 1.6581 0.3304 1.6581 1.2877
No log 3.0769 40 1.6237 0.3130 1.6237 1.2743
No log 3.2308 42 1.6067 0.2807 1.6067 1.2676
No log 3.3846 44 1.7004 0.2281 1.7004 1.3040
No log 3.5385 46 2.0071 0.1679 2.0071 1.4167
No log 3.6923 48 1.9563 0.2000 1.9563 1.3987
No log 3.8462 50 1.8531 0.2419 1.8531 1.3613
No log 4.0 52 1.5111 0.3607 1.5111 1.2293
No log 4.1538 54 1.3889 0.3607 1.3889 1.1785
No log 4.3077 56 1.2872 0.4 1.2872 1.1345
No log 4.4615 58 1.5845 0.3214 1.5845 1.2588
No log 4.6154 60 1.8408 0.2609 1.8408 1.3568
No log 4.7692 62 1.8535 0.125 1.8535 1.3615
No log 4.9231 64 1.6437 0.2051 1.6437 1.2821
No log 5.0769 66 1.7009 0.2069 1.7009 1.3042
No log 5.2308 68 1.6856 0.2881 1.6856 1.2983
No log 5.3846 70 1.9165 0.2167 1.9165 1.3844
No log 5.5385 72 1.5649 0.3279 1.5649 1.2509
No log 5.6923 74 1.2928 0.3968 1.2928 1.1370
No log 5.8462 76 1.2650 0.4375 1.2650 1.1247
No log 6.0 78 1.2150 0.4531 1.2150 1.1023
No log 6.1538 80 1.2586 0.4531 1.2586 1.1219
No log 6.3077 82 1.4548 0.4211 1.4548 1.2062
No log 6.4615 84 1.5649 0.3704 1.5649 1.2510
No log 6.6154 86 1.3087 0.5152 1.3087 1.1440
No log 6.7692 88 1.0822 0.5588 1.0822 1.0403
No log 6.9231 90 1.0833 0.6187 1.0833 1.0408
No log 7.0769 92 1.4087 0.3577 1.4087 1.1869
No log 7.2308 94 1.7925 0.2333 1.7925 1.3388
No log 7.3846 96 1.3935 0.4308 1.3935 1.1805
No log 7.5385 98 0.9774 0.6099 0.9774 0.9886
No log 7.6923 100 1.1058 0.6259 1.1058 1.0516
No log 7.8462 102 1.0656 0.6479 1.0656 1.0323
No log 8.0 104 0.9163 0.6471 0.9163 0.9573
No log 8.1538 106 1.3513 0.4412 1.3513 1.1625
No log 8.3077 108 1.7677 0.3429 1.7677 1.3296
No log 8.4615 110 1.6991 0.3623 1.6991 1.3035
No log 8.6154 112 1.3392 0.5191 1.3392 1.1572
No log 8.7692 114 1.1636 0.528 1.1636 1.0787
No log 8.9231 116 1.1739 0.5909 1.1739 1.0835
No log 9.0769 118 1.1446 0.5802 1.1446 1.0698
No log 9.2308 120 1.2349 0.544 1.2349 1.1112
No log 9.3846 122 1.4846 0.4567 1.4846 1.2184
No log 9.5385 124 1.4936 0.4478 1.4936 1.2221
No log 9.6923 126 1.2635 0.5303 1.2635 1.1240
No log 9.8462 128 1.0869 0.5714 1.0869 1.0425
No log 10.0 130 0.9761 0.5846 0.9761 0.9880
No log 10.1538 132 0.9804 0.5581 0.9804 0.9902
No log 10.3077 134 1.1084 0.5891 1.1084 1.0528
No log 10.4615 136 1.2538 0.5191 1.2538 1.1197
No log 10.6154 138 1.2168 0.5469 1.2168 1.1031
No log 10.7692 140 1.0818 0.5625 1.0818 1.0401
No log 10.9231 142 1.0205 0.5581 1.0205 1.0102
No log 11.0769 144 1.0249 0.5538 1.0249 1.0124
No log 11.2308 146 1.1427 0.5606 1.1427 1.0690
No log 11.3846 148 1.2567 0.5303 1.2567 1.1210
No log 11.5385 150 1.2181 0.5564 1.2181 1.1037
No log 11.6923 152 1.2054 0.5672 1.2054 1.0979
No log 11.8462 154 1.3563 0.5113 1.3563 1.1646
No log 12.0 156 1.2386 0.5191 1.2386 1.1129
No log 12.1538 158 1.1867 0.5231 1.1867 1.0894
No log 12.3077 160 1.2365 0.4923 1.2365 1.1120
No log 12.4615 162 1.3454 0.5075 1.3454 1.1599
No log 12.6154 164 1.3618 0.5075 1.3618 1.1669
No log 12.7692 166 1.2806 0.5113 1.2806 1.1316
No log 12.9231 168 1.0414 0.496 1.0414 1.0205
No log 13.0769 170 0.9196 0.6765 0.9196 0.9590
No log 13.2308 172 0.8923 0.6765 0.8923 0.9446
No log 13.3846 174 0.8277 0.6765 0.8277 0.9098
No log 13.5385 176 0.8631 0.6715 0.8631 0.9290
No log 13.6923 178 1.0513 0.6074 1.0513 1.0253
No log 13.8462 180 1.1146 0.5564 1.1146 1.0557
No log 14.0 182 1.0165 0.6119 1.0165 1.0082
No log 14.1538 184 0.9229 0.6269 0.9229 0.9607
No log 14.3077 186 0.9040 0.6324 0.9040 0.9508
No log 14.4615 188 0.9641 0.5496 0.9641 0.9819
No log 14.6154 190 1.1484 0.4961 1.1484 1.0717
No log 14.7692 192 1.3804 0.4962 1.3804 1.1749
No log 14.9231 194 1.4432 0.5 1.4432 1.2013
No log 15.0769 196 1.4312 0.5 1.4312 1.1963
No log 15.2308 198 1.3359 0.4640 1.3359 1.1558
No log 15.3846 200 1.2920 0.4640 1.2920 1.1367
No log 15.5385 202 1.2774 0.4640 1.2774 1.1302
No log 15.6923 204 1.3782 0.4733 1.3782 1.1739
No log 15.8462 206 1.3919 0.4361 1.3919 1.1798
No log 16.0 208 1.2452 0.4806 1.2452 1.1159
No log 16.1538 210 1.0929 0.496 1.0929 1.0454
No log 16.3077 212 1.0251 0.5802 1.0251 1.0125
No log 16.4615 214 1.0155 0.5354 1.0155 1.0077
No log 16.6154 216 1.1139 0.4762 1.1139 1.0554
No log 16.7692 218 1.3278 0.5152 1.3278 1.1523
No log 16.9231 220 1.3762 0.5113 1.3762 1.1731
No log 17.0769 222 1.2889 0.5038 1.2889 1.1353
No log 17.2308 224 1.1581 0.4961 1.1581 1.0762
No log 17.3846 226 1.0620 0.5344 1.0620 1.0305
No log 17.5385 228 1.0996 0.4733 1.0996 1.0486
No log 17.6923 230 1.2433 0.4961 1.2433 1.1150
No log 17.8462 232 1.4066 0.4511 1.4066 1.1860
No log 18.0 234 1.4382 0.4296 1.4382 1.1992
No log 18.1538 236 1.2966 0.5038 1.2966 1.1387
No log 18.3077 238 1.1749 0.5231 1.1749 1.0839
No log 18.4615 240 1.1491 0.5161 1.1491 1.0720
No log 18.6154 242 1.1566 0.5714 1.1566 1.0755
No log 18.7692 244 1.0805 0.5556 1.0805 1.0395
No log 18.9231 246 0.9961 0.5625 0.9961 0.9981
No log 19.0769 248 0.9789 0.5781 0.9789 0.9894
No log 19.2308 250 1.0130 0.5354 1.0130 1.0065
No log 19.3846 252 1.0693 0.5512 1.0693 1.0341
No log 19.5385 254 1.1972 0.5231 1.1972 1.0942
No log 19.6923 256 1.1618 0.5581 1.1618 1.0779
No log 19.8462 258 1.0534 0.512 1.0534 1.0264
No log 20.0 260 1.0353 0.512 1.0353 1.0175
No log 20.1538 262 1.1296 0.5354 1.1296 1.0628
No log 20.3077 264 1.3255 0.5231 1.3255 1.1513
No log 20.4615 266 1.4132 0.5 1.4132 1.1888
No log 20.6154 268 1.3496 0.5231 1.3496 1.1617
No log 20.7692 270 1.2807 0.5231 1.2807 1.1317
No log 20.9231 272 1.3154 0.5231 1.3154 1.1469
No log 21.0769 274 1.3577 0.5 1.3577 1.1652
No log 21.2308 276 1.3949 0.4812 1.3949 1.1810
No log 21.3846 278 1.3912 0.4812 1.3912 1.1795
No log 21.5385 280 1.2992 0.5581 1.2992 1.1398
No log 21.6923 282 1.2664 0.5469 1.2664 1.1253
No log 21.8462 284 1.1921 0.5197 1.1921 1.0918
No log 22.0 286 1.0830 0.5238 1.0830 1.0407
No log 22.1538 288 0.9814 0.6165 0.9814 0.9906
No log 22.3077 290 0.9587 0.6418 0.9587 0.9791
No log 22.4615 292 0.9954 0.6324 0.9954 0.9977
No log 22.6154 294 1.0653 0.6029 1.0653 1.0322
No log 22.7692 296 0.9465 0.6423 0.9465 0.9729
No log 22.9231 298 0.8905 0.6370 0.8905 0.9437
No log 23.0769 300 0.9172 0.6212 0.9172 0.9577
No log 23.2308 302 0.9770 0.5556 0.9770 0.9884
No log 23.3846 304 1.0801 0.4386 1.0801 1.0393
No log 23.5385 306 1.1917 0.4174 1.1917 1.0917
No log 23.6923 308 1.3068 0.5289 1.3068 1.1432
No log 23.8462 310 1.3507 0.5156 1.3507 1.1622
No log 24.0 312 1.2848 0.5496 1.2848 1.1335
No log 24.1538 314 1.1512 0.5271 1.1512 1.0729
No log 24.3077 316 1.0043 0.5079 1.0043 1.0022
No log 24.4615 318 0.9562 0.5802 0.9562 0.9778
No log 24.6154 320 0.9387 0.5758 0.9387 0.9689
No log 24.7692 322 1.0004 0.5469 1.0004 1.0002
No log 24.9231 324 1.1808 0.5496 1.1808 1.0867
No log 25.0769 326 1.4417 0.4627 1.4417 1.2007
No log 25.2308 328 1.5235 0.4179 1.5235 1.2343
No log 25.3846 330 1.4478 0.5077 1.4478 1.2033
No log 25.5385 332 1.2698 0.5538 1.2698 1.1269
No log 25.6923 334 1.0958 0.5354 1.0958 1.0468
No log 25.8462 336 1.0489 0.5397 1.0489 1.0242
No log 26.0 338 1.0128 0.5512 1.0128 1.0064
No log 26.1538 340 1.0206 0.5397 1.0206 1.0103
No log 26.3077 342 1.0658 0.512 1.0658 1.0324
No log 26.4615 344 1.1359 0.5397 1.1359 1.0658
No log 26.6154 346 1.2313 0.5581 1.2313 1.1096
No log 26.7692 348 1.2564 0.5538 1.2564 1.1209
No log 26.9231 350 1.2757 0.5538 1.2757 1.1295
No log 27.0769 352 1.2956 0.5538 1.2956 1.1382
No log 27.2308 354 1.3668 0.5538 1.3668 1.1691
No log 27.3846 356 1.3165 0.5538 1.3165 1.1474
No log 27.5385 358 1.2688 0.5581 1.2688 1.1264
No log 27.6923 360 1.2316 0.5581 1.2316 1.1098
No log 27.8462 362 1.2121 0.496 1.2121 1.1010
No log 28.0 364 1.2512 0.496 1.2512 1.1186
No log 28.1538 366 1.3372 0.5354 1.3372 1.1564
No log 28.3077 368 1.4182 0.5039 1.4182 1.1909
No log 28.4615 370 1.4922 0.4091 1.4922 1.2216
No log 28.6154 372 1.4768 0.4462 1.4768 1.2152
No log 28.7692 374 1.4202 0.4688 1.4202 1.1917
No log 28.9231 376 1.3774 0.4961 1.3774 1.1736
No log 29.0769 378 1.2642 0.5581 1.2642 1.1244
No log 29.2308 380 1.1794 0.5312 1.1794 1.0860
No log 29.3846 382 1.1172 0.4677 1.1172 1.0570
No log 29.5385 384 1.1090 0.4590 1.1090 1.0531
No log 29.6923 386 1.1314 0.4590 1.1314 1.0637
No log 29.8462 388 1.2214 0.5354 1.2214 1.1052
No log 30.0 390 1.3783 0.5 1.3783 1.1740
No log 30.1538 392 1.4919 0.4091 1.4919 1.2215
No log 30.3077 394 1.5016 0.4091 1.5016 1.2254
No log 30.4615 396 1.4179 0.4806 1.4179 1.1908
No log 30.6154 398 1.3117 0.5238 1.3117 1.1453
No log 30.7692 400 1.2296 0.528 1.2296 1.1089
No log 30.9231 402 1.2120 0.5 1.2120 1.1009
No log 31.0769 404 1.2407 0.528 1.2407 1.1139
No log 31.2308 406 1.2843 0.528 1.2843 1.1333
No log 31.3846 408 1.3329 0.4677 1.3329 1.1545
No log 31.5385 410 1.3428 0.48 1.3428 1.1588
No log 31.6923 412 1.3532 0.5039 1.3532 1.1633
No log 31.8462 414 1.3355 0.5039 1.3355 1.1556
No log 32.0 416 1.2802 0.5039 1.2802 1.1314
No log 32.1538 418 1.2318 0.5312 1.2318 1.1099
No log 32.3077 420 1.1497 0.528 1.1497 1.0722
No log 32.4615 422 1.1105 0.528 1.1105 1.0538
No log 32.6154 424 1.0610 0.4463 1.0610 1.0301
No log 32.7692 426 1.0552 0.4878 1.0552 1.0272
No log 32.9231 428 1.0753 0.4615 1.0753 1.0370
No log 33.0769 430 1.1189 0.4500 1.1189 1.0578
No log 33.2308 432 1.2114 0.496 1.2114 1.1006
No log 33.3846 434 1.2940 0.5512 1.2940 1.1375
No log 33.5385 436 1.3529 0.5581 1.3529 1.1631
No log 33.6923 438 1.3442 0.5581 1.3442 1.1594
No log 33.8462 440 1.2989 0.5312 1.2989 1.1397
No log 34.0 442 1.2979 0.5312 1.2979 1.1393
No log 34.1538 444 1.2947 0.4640 1.2947 1.1378
No log 34.3077 446 1.2503 0.4516 1.2503 1.1182
No log 34.4615 448 1.2261 0.4754 1.2261 1.1073
No log 34.6154 450 1.2378 0.4754 1.2378 1.1126
No log 34.7692 452 1.2733 0.4839 1.2733 1.1284
No log 34.9231 454 1.3385 0.5039 1.3385 1.1570
No log 35.0769 456 1.3576 0.5312 1.3576 1.1651
No log 35.2308 458 1.3506 0.5312 1.3506 1.1622
No log 35.3846 460 1.3800 0.5312 1.3800 1.1747
No log 35.5385 462 1.3415 0.5312 1.3415 1.1582
No log 35.6923 464 1.2825 0.4640 1.2825 1.1325
No log 35.8462 466 1.2715 0.4640 1.2715 1.1276
No log 36.0 468 1.2631 0.4640 1.2631 1.1239
No log 36.1538 470 1.2755 0.5197 1.2755 1.1294
No log 36.3077 472 1.2788 0.5197 1.2788 1.1308
No log 36.4615 474 1.2547 0.5197 1.2547 1.1201
No log 36.6154 476 1.1975 0.4921 1.1975 1.0943
No log 36.7692 478 1.1341 0.4640 1.1341 1.0649
No log 36.9231 480 1.1434 0.5039 1.1434 1.0693
No log 37.0769 482 1.1592 0.5039 1.1592 1.0766
No log 37.2308 484 1.1932 0.5039 1.1932 1.0923
No log 37.3846 486 1.2268 0.5039 1.2268 1.1076
No log 37.5385 488 1.2459 0.48 1.2459 1.1162
No log 37.6923 490 1.2417 0.4553 1.2417 1.1143
No log 37.8462 492 1.2230 0.4553 1.2230 1.1059
No log 38.0 494 1.2163 0.4553 1.2163 1.1029
No log 38.1538 496 1.2647 0.5079 1.2647 1.1246
No log 38.3077 498 1.3363 0.5581 1.3363 1.1560
0.3462 38.4615 500 1.3440 0.5496 1.3440 1.1593
0.3462 38.6154 502 1.3042 0.5581 1.3042 1.1420
0.3462 38.7692 504 1.2260 0.5312 1.2260 1.1073
0.3462 38.9231 506 1.1640 0.5197 1.1640 1.0789
0.3462 39.0769 508 1.1464 0.5197 1.1464 1.0707
0.3462 39.2308 510 1.1342 0.4516 1.1342 1.0650
0.3462 39.3846 512 1.1380 0.4516 1.1380 1.0668

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k3_task1_organization

Finetuned
(4019)
this model