ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9668
  • Qwk: 0.6107
  • Mse: 0.9668
  • Rmse: 0.9832

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0333 2 6.8758 0.0239 6.8758 2.6222
No log 0.0667 4 4.4974 0.0263 4.4974 2.1207
No log 0.1 6 3.2825 0.0 3.2825 1.8118
No log 0.1333 8 2.3423 0.1127 2.3423 1.5304
No log 0.1667 10 2.2833 0.1871 2.2833 1.5110
No log 0.2 12 2.0649 0.2185 2.0649 1.4370
No log 0.2333 14 1.7901 0.1770 1.7901 1.3379
No log 0.2667 16 1.8344 0.1964 1.8344 1.3544
No log 0.3 18 1.8191 0.1947 1.8191 1.3488
No log 0.3333 20 1.6403 0.1636 1.6403 1.2807
No log 0.3667 22 1.6875 0.1869 1.6875 1.2990
No log 0.4 24 1.6601 0.2342 1.6601 1.2884
No log 0.4333 26 2.0253 0.2774 2.0253 1.4231
No log 0.4667 28 3.2665 0.0978 3.2665 1.8074
No log 0.5 30 4.1655 0.1111 4.1655 2.0410
No log 0.5333 32 3.4058 0.1451 3.4058 1.8455
No log 0.5667 34 2.0291 0.2993 2.0291 1.4245
No log 0.6 36 1.5327 0.2342 1.5327 1.2380
No log 0.6333 38 1.4607 0.2018 1.4607 1.2086
No log 0.6667 40 1.5096 0.2182 1.5096 1.2287
No log 0.7 42 1.7315 0.35 1.7315 1.3159
No log 0.7333 44 1.8449 0.2576 1.8449 1.3583
No log 0.7667 46 2.0307 0.3143 2.0307 1.4250
No log 0.8 48 1.8582 0.3099 1.8582 1.3632
No log 0.8333 50 1.4690 0.4154 1.4690 1.2120
No log 0.8667 52 1.2919 0.4688 1.2919 1.1366
No log 0.9 54 1.2985 0.4427 1.2985 1.1395
No log 0.9333 56 1.3202 0.4662 1.3202 1.1490
No log 0.9667 58 1.3076 0.4545 1.3076 1.1435
No log 1.0 60 1.2653 0.3871 1.2653 1.1249
No log 1.0333 62 1.3072 0.2783 1.3072 1.1433
No log 1.0667 64 1.3595 0.2857 1.3595 1.1660
No log 1.1 66 1.5133 0.3676 1.5133 1.2302
No log 1.1333 68 1.9712 0.3038 1.9712 1.4040
No log 1.1667 70 1.7246 0.4138 1.7246 1.3132
No log 1.2 72 1.3395 0.4697 1.3395 1.1574
No log 1.2333 74 1.4342 0.4480 1.4342 1.1976
No log 1.2667 76 1.3229 0.5312 1.3229 1.1502
No log 1.3 78 1.2920 0.4921 1.2920 1.1367
No log 1.3333 80 1.3557 0.3206 1.3557 1.1643
No log 1.3667 82 1.2656 0.4275 1.2656 1.1250
No log 1.4 84 1.1883 0.5231 1.1883 1.0901
No log 1.4333 86 1.1873 0.5116 1.1873 1.0896
No log 1.4667 88 1.2589 0.5116 1.2589 1.1220
No log 1.5 90 1.3995 0.4545 1.3995 1.1830
No log 1.5333 92 1.1728 0.5231 1.1728 1.0830
No log 1.5667 94 1.1583 0.5970 1.1583 1.0762
No log 1.6 96 1.1930 0.6107 1.1930 1.0922
No log 1.6333 98 1.1150 0.5758 1.1150 1.0559
No log 1.6667 100 1.1801 0.5255 1.1801 1.0863
No log 1.7 102 1.2684 0.5143 1.2684 1.1262
No log 1.7333 104 1.1843 0.5037 1.1843 1.0883
No log 1.7667 106 1.1362 0.5984 1.1362 1.0659
No log 1.8 108 1.5879 0.3025 1.5879 1.2601
No log 1.8333 110 1.8080 0.2609 1.8080 1.3446
No log 1.8667 112 1.5459 0.3051 1.5459 1.2433
No log 1.9 114 1.1820 0.528 1.1820 1.0872
No log 1.9333 116 1.0987 0.4640 1.0987 1.0482
No log 1.9667 118 1.1747 0.5571 1.1747 1.0838
No log 2.0 120 1.1430 0.5294 1.1430 1.0691
No log 2.0333 122 1.1738 0.48 1.1738 1.0834
No log 2.0667 124 1.2961 0.4262 1.2961 1.1385
No log 2.1 126 1.2289 0.5248 1.2289 1.1086
No log 2.1333 128 1.1354 0.5417 1.1354 1.0656
No log 2.1667 130 1.1968 0.5850 1.1968 1.0940
No log 2.2 132 1.1144 0.5986 1.1144 1.0556
No log 2.2333 134 1.1801 0.5753 1.1801 1.0863
No log 2.2667 136 1.2059 0.5390 1.2059 1.0981
No log 2.3 138 1.2329 0.4531 1.2329 1.1104
No log 2.3333 140 1.4307 0.5077 1.4307 1.1961
No log 2.3667 142 1.3787 0.5039 1.3787 1.1742
No log 2.4 144 1.2044 0.4603 1.2044 1.0974
No log 2.4333 146 1.2988 0.5612 1.2988 1.1397
No log 2.4667 148 1.3409 0.4478 1.3409 1.1580
No log 2.5 150 1.2704 0.5185 1.2704 1.1271
No log 2.5333 152 1.1829 0.5802 1.1829 1.0876
No log 2.5667 154 1.2108 0.4715 1.2108 1.1004
No log 2.6 156 1.3885 0.5116 1.3885 1.1784
No log 2.6333 158 1.4441 0.5369 1.4441 1.2017
No log 2.6667 160 1.1946 0.5333 1.1946 1.0930
No log 2.7 162 1.0468 0.5581 1.0468 1.0231
No log 2.7333 164 1.1251 0.6 1.1251 1.0607
No log 2.7667 166 1.1847 0.6 1.1847 1.0884
No log 2.8 168 1.1436 0.5588 1.1436 1.0694
No log 2.8333 170 1.0915 0.6131 1.0915 1.0447
No log 2.8667 172 1.1302 0.5594 1.1302 1.0631
No log 2.9 174 1.1703 0.4892 1.1703 1.0818
No log 2.9333 176 1.1975 0.5224 1.1975 1.0943
No log 2.9667 178 1.1650 0.4132 1.1650 1.0793
No log 3.0 180 1.1687 0.4034 1.1687 1.0810
No log 3.0333 182 1.1704 0.4628 1.1704 1.0819
No log 3.0667 184 1.1731 0.5354 1.1731 1.0831
No log 3.1 186 1.1115 0.5556 1.1115 1.0543
No log 3.1333 188 1.0912 0.544 1.0912 1.0446
No log 3.1667 190 1.0963 0.4959 1.0963 1.0470
No log 3.2 192 1.1131 0.4237 1.1131 1.0550
No log 3.2333 194 1.1380 0.3621 1.1380 1.0668
No log 3.2667 196 1.1671 0.3621 1.1671 1.0803
No log 3.3 198 1.1807 0.3932 1.1807 1.0866
No log 3.3333 200 1.1808 0.3697 1.1808 1.0866
No log 3.3667 202 1.1878 0.5547 1.1878 1.0898
No log 3.4 204 1.2035 0.5429 1.2035 1.0970
No log 3.4333 206 1.1179 0.5802 1.1179 1.0573
No log 3.4667 208 1.2094 0.5362 1.2094 1.0997
No log 3.5 210 1.4737 0.4605 1.4737 1.2140
No log 3.5333 212 1.2544 0.5211 1.2544 1.1200
No log 3.5667 214 1.0503 0.5970 1.0503 1.0249
No log 3.6 216 1.0657 0.6074 1.0657 1.0324
No log 3.6333 218 1.0614 0.6212 1.0614 1.0302
No log 3.6667 220 1.0187 0.5736 1.0187 1.0093
No log 3.7 222 1.0129 0.5564 1.0129 1.0064
No log 3.7333 224 0.9930 0.5496 0.9930 0.9965
No log 3.7667 226 0.9919 0.5736 0.9919 0.9960
No log 3.8 228 1.0006 0.5581 1.0006 1.0003
No log 3.8333 230 1.0182 0.5846 1.0182 1.0091
No log 3.8667 232 1.0354 0.5426 1.0354 1.0176
No log 3.9 234 1.0488 0.5865 1.0488 1.0241
No log 3.9333 236 1.0599 0.5692 1.0599 1.0295
No log 3.9667 238 1.0565 0.5 1.0565 1.0279
No log 4.0 240 1.0708 0.5039 1.0708 1.0348
No log 4.0333 242 1.0724 0.48 1.0724 1.0356
No log 4.0667 244 1.0683 0.5238 1.0683 1.0336
No log 4.1 246 1.0546 0.5692 1.0546 1.0269
No log 4.1333 248 0.9791 0.6119 0.9791 0.9895
No log 4.1667 250 0.9497 0.6471 0.9497 0.9745
No log 4.2 252 0.9922 0.5797 0.9922 0.9961
No log 4.2333 254 1.0595 0.5899 1.0595 1.0293
No log 4.2667 256 1.0125 0.5758 1.0125 1.0062
No log 4.3 258 0.9614 0.528 0.9614 0.9805
No log 4.3333 260 0.9600 0.5781 0.9600 0.9798
No log 4.3667 262 0.9675 0.6047 0.9675 0.9836
No log 4.4 264 0.9537 0.6047 0.9537 0.9766
No log 4.4333 266 0.9326 0.6222 0.9326 0.9657
No log 4.4667 268 0.9624 0.6143 0.9624 0.9810
No log 4.5 270 1.0838 0.5694 1.0838 1.0410
No log 4.5333 272 0.9861 0.5899 0.9861 0.9930
No log 4.5667 274 0.9748 0.5496 0.9748 0.9873
No log 4.6 276 1.0713 0.5 1.0713 1.0351
No log 4.6333 278 1.0992 0.5 1.0992 1.0484
No log 4.6667 280 1.0646 0.5079 1.0646 1.0318
No log 4.7 282 0.9969 0.5079 0.9969 0.9985
No log 4.7333 284 0.9820 0.5693 0.9820 0.9910
No log 4.7667 286 0.9990 0.5816 0.9990 0.9995
No log 4.8 288 0.9372 0.6528 0.9372 0.9681
No log 4.8333 290 0.8785 0.7027 0.8785 0.9373
No log 4.8667 292 0.8783 0.7123 0.8783 0.9372
No log 4.9 294 0.9095 0.6519 0.9095 0.9537
No log 4.9333 296 0.8910 0.6765 0.8910 0.9439
No log 4.9667 298 0.8981 0.6277 0.8981 0.9477
No log 5.0 300 0.9370 0.5455 0.9370 0.9680
No log 5.0333 302 0.9029 0.6074 0.9029 0.9502
No log 5.0667 304 0.8472 0.7 0.8472 0.9204
No log 5.1 306 0.8210 0.7260 0.8210 0.9061
No log 5.1333 308 0.8103 0.7162 0.8103 0.9002
No log 5.1667 310 0.8150 0.7248 0.8150 0.9028
No log 5.2 312 0.8527 0.6757 0.8527 0.9234
No log 5.2333 314 0.8652 0.6713 0.8652 0.9301
No log 5.2667 316 0.8798 0.6713 0.8798 0.9380
No log 5.3 318 0.8436 0.7172 0.8436 0.9185
No log 5.3333 320 0.8318 0.72 0.8318 0.9120
No log 5.3667 322 0.8842 0.6980 0.8842 0.9403
No log 5.4 324 0.8822 0.7162 0.8822 0.9392
No log 5.4333 326 0.8089 0.7383 0.8089 0.8994
No log 5.4667 328 0.8025 0.7260 0.8025 0.8958
No log 5.5 330 0.8062 0.7211 0.8062 0.8979
No log 5.5333 332 0.8455 0.7042 0.8455 0.9195
No log 5.5667 334 0.8485 0.6906 0.8485 0.9212
No log 5.6 336 0.8850 0.6667 0.8850 0.9408
No log 5.6333 338 0.9229 0.5672 0.9229 0.9607
No log 5.6667 340 0.9731 0.5522 0.9731 0.9865
No log 5.7 342 0.9374 0.5942 0.9374 0.9682
No log 5.7333 344 0.8631 0.6906 0.8631 0.9290
No log 5.7667 346 0.8558 0.7050 0.8558 0.9251
No log 5.8 348 0.8734 0.7 0.8734 0.9345
No log 5.8333 350 0.9152 0.6316 0.9152 0.9567
No log 5.8667 352 0.9131 0.6316 0.9131 0.9555
No log 5.9 354 0.9986 0.5303 0.9986 0.9993
No log 5.9333 356 1.1283 0.5390 1.1283 1.0622
No log 5.9667 358 1.1288 0.5507 1.1288 1.0625
No log 6.0 360 1.0300 0.4844 1.0300 1.0149
No log 6.0333 362 0.9636 0.528 0.9636 0.9816
No log 6.0667 364 0.9757 0.6412 0.9757 0.9878
No log 6.1 366 0.9712 0.6418 0.9712 0.9855
No log 6.1333 368 0.9490 0.6466 0.9490 0.9742
No log 6.1667 370 0.9399 0.5938 0.9399 0.9695
No log 6.2 372 0.9385 0.6202 0.9385 0.9688
No log 6.2333 374 0.9395 0.5909 0.9395 0.9693
No log 6.2667 376 0.9347 0.6107 0.9347 0.9668
No log 6.3 378 0.9324 0.6202 0.9324 0.9656
No log 6.3333 380 0.9407 0.6418 0.9407 0.9699
No log 6.3667 382 0.9069 0.6269 0.9069 0.9523
No log 6.4 384 0.8748 0.7183 0.8748 0.9353
No log 6.4333 386 0.9158 0.6423 0.9158 0.9570
No log 6.4667 388 0.9247 0.6015 0.9247 0.9616
No log 6.5 390 0.9111 0.6861 0.9111 0.9545
No log 6.5333 392 0.8973 0.6861 0.8973 0.9472
No log 6.5667 394 0.8979 0.6957 0.8979 0.9476
No log 6.6 396 0.8913 0.6715 0.8913 0.9441
No log 6.6333 398 0.9211 0.6 0.9211 0.9597
No log 6.6667 400 0.9594 0.496 0.9594 0.9795
No log 6.7 402 0.9794 0.56 0.9794 0.9897
No log 6.7333 404 0.9869 0.5714 0.9869 0.9934
No log 6.7667 406 0.9475 0.6412 0.9475 0.9734
No log 6.8 408 0.9545 0.6617 0.9545 0.9770
No log 6.8333 410 0.9035 0.6466 0.9035 0.9505
No log 6.8667 412 0.8966 0.6269 0.8966 0.9469
No log 6.9 414 0.9029 0.5909 0.9029 0.9502
No log 6.9333 416 0.9166 0.5736 0.9166 0.9574
No log 6.9667 418 0.9169 0.6 0.9169 0.9575
No log 7.0 420 0.8751 0.6269 0.8751 0.9354
No log 7.0333 422 0.8468 0.6986 0.8468 0.9202
No log 7.0667 424 0.8175 0.7075 0.8175 0.9042
No log 7.1 426 0.8040 0.6986 0.8040 0.8967
No log 7.1333 428 0.8708 0.6809 0.8708 0.9332
No log 7.1667 430 0.8933 0.6667 0.8933 0.9451
No log 7.2 432 0.8719 0.6667 0.8719 0.9338
No log 7.2333 434 0.9390 0.6056 0.9390 0.9690
No log 7.2667 436 1.0037 0.5775 1.0037 1.0019
No log 7.3 438 1.0097 0.5414 1.0097 1.0048
No log 7.3333 440 0.9999 0.4463 0.9999 0.9999
No log 7.3667 442 1.0100 0.5 1.0100 1.0050
No log 7.4 444 1.0248 0.4754 1.0248 1.0123
No log 7.4333 446 1.0228 0.4754 1.0228 1.0113
No log 7.4667 448 1.0210 0.4754 1.0210 1.0105
No log 7.5 450 0.9887 0.5079 0.9887 0.9943
No log 7.5333 452 0.9733 0.4878 0.9733 0.9865
No log 7.5667 454 0.9719 0.5556 0.9719 0.9858
No log 7.6 456 1.0131 0.528 1.0131 1.0065
No log 7.6333 458 1.0254 0.528 1.0254 1.0126
No log 7.6667 460 0.9931 0.5556 0.9931 0.9966
No log 7.7 462 1.0569 0.5 1.0569 1.0281
No log 7.7333 464 1.1855 0.5899 1.1855 1.0888
No log 7.7667 466 1.2280 0.5867 1.2280 1.1081
No log 7.8 468 1.1265 0.6013 1.1265 1.0614
No log 7.8333 470 0.9237 0.5753 0.9237 0.9611
No log 7.8667 472 0.8054 0.7347 0.8054 0.8975
No log 7.9 474 0.8056 0.6957 0.8056 0.8976
No log 7.9333 476 0.8147 0.6716 0.8147 0.9026
No log 7.9667 478 0.8438 0.6667 0.8438 0.9186
No log 8.0 480 0.8883 0.6767 0.8883 0.9425
No log 8.0333 482 0.9168 0.6154 0.9168 0.9575
No log 8.0667 484 0.9244 0.6364 0.9244 0.9615
No log 8.1 486 0.9153 0.6466 0.9153 0.9567
No log 8.1333 488 0.8996 0.6515 0.8996 0.9485
No log 8.1667 490 0.9209 0.6515 0.9209 0.9596
No log 8.2 492 0.9618 0.6 0.9618 0.9807
No log 8.2333 494 0.9600 0.6260 0.9600 0.9798
No log 8.2667 496 0.9355 0.6471 0.9355 0.9672
No log 8.3 498 0.9492 0.6569 0.9492 0.9743
0.3973 8.3333 500 0.9255 0.6569 0.9255 0.9620
0.3973 8.3667 502 0.9048 0.6418 0.9048 0.9512
0.3973 8.4 504 0.9130 0.6316 0.9130 0.9555
0.3973 8.4333 506 0.9345 0.6316 0.9345 0.9667
0.3973 8.4667 508 0.9649 0.5954 0.9649 0.9823
0.3973 8.5 510 0.9668 0.6107 0.9668 0.9832

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k13_task1_organization

Finetuned
(4019)
this model