ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k12_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8079
  • Qwk: 0.6301
  • Mse: 0.8079
  • Rmse: 0.8988

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0351 2 6.9296 0.0242 6.9296 2.6324
No log 0.0702 4 4.7337 0.0821 4.7337 2.1757
No log 0.1053 6 3.6823 -0.0104 3.6823 1.9189
No log 0.1404 8 2.6693 0.1325 2.6693 1.6338
No log 0.1754 10 2.2112 0.0945 2.2112 1.4870
No log 0.2105 12 2.0987 0.1322 2.0987 1.4487
No log 0.2456 14 2.5361 -0.0272 2.5361 1.5925
No log 0.2807 16 3.0326 0.0242 3.0326 1.7414
No log 0.3158 18 2.9031 0.0252 2.9031 1.7039
No log 0.3509 20 2.2513 0.0597 2.2513 1.5004
No log 0.3860 22 2.1247 0.0630 2.1247 1.4576
No log 0.4211 24 2.0723 0.1692 2.0723 1.4396
No log 0.4561 26 2.0352 0.2000 2.0352 1.4266
No log 0.4912 28 2.0163 0.1940 2.0163 1.4200
No log 0.5263 30 2.0028 0.2388 2.0028 1.4152
No log 0.5614 32 2.2032 0.2069 2.2032 1.4843
No log 0.5965 34 2.1125 0.2361 2.1125 1.4534
No log 0.6316 36 1.8307 0.3165 1.8307 1.3530
No log 0.6667 38 2.0111 0.2877 2.0111 1.4181
No log 0.7018 40 3.0686 0.1413 3.0686 1.7518
No log 0.7368 42 4.3040 0.0775 4.3040 2.0746
No log 0.7719 44 4.5569 0.0815 4.5569 2.1347
No log 0.8070 46 3.9377 0.1795 3.9377 1.9844
No log 0.8421 48 2.8113 0.1420 2.8113 1.6767
No log 0.8772 50 1.7331 0.3382 1.7331 1.3165
No log 0.9123 52 1.3927 0.3871 1.3927 1.1801
No log 0.9474 54 1.4406 0.3871 1.4406 1.2002
No log 0.9825 56 1.4921 0.3846 1.4921 1.2215
No log 1.0175 58 1.7681 0.3636 1.7681 1.3297
No log 1.0526 60 2.0518 0.3108 2.0518 1.4324
No log 1.0877 62 2.5116 0.1829 2.5116 1.5848
No log 1.1228 64 3.0561 0.2451 3.0561 1.7482
No log 1.1579 66 3.0566 0.2388 3.0566 1.7483
No log 1.1930 68 2.7300 0.1932 2.7300 1.6523
No log 1.2281 70 2.2474 0.2025 2.2474 1.4991
No log 1.2632 72 1.6316 0.4366 1.6316 1.2773
No log 1.2982 74 1.2980 0.5075 1.2980 1.1393
No log 1.3333 76 1.2123 0.4812 1.2123 1.1011
No log 1.3684 78 1.3066 0.4662 1.3066 1.1431
No log 1.4035 80 1.6809 0.4552 1.6809 1.2965
No log 1.4386 82 2.5542 0.2985 2.5542 1.5982
No log 1.4737 84 2.9914 0.2453 2.9914 1.7296
No log 1.5088 86 2.7429 0.2549 2.7429 1.6562
No log 1.5439 88 2.1311 0.3699 2.1311 1.4598
No log 1.5789 90 1.8040 0.4113 1.8040 1.3431
No log 1.6140 92 1.6314 0.4604 1.6314 1.2773
No log 1.6491 94 1.4567 0.4672 1.4567 1.2069
No log 1.6842 96 1.4532 0.4857 1.4532 1.2055
No log 1.7193 98 1.7790 0.3946 1.7790 1.3338
No log 1.7544 100 1.9188 0.4192 1.9188 1.3852
No log 1.7895 102 2.3889 0.3420 2.3889 1.5456
No log 1.8246 104 3.1674 0.2566 3.1674 1.7797
No log 1.8596 106 3.0849 0.2566 3.0849 1.7564
No log 1.8947 108 2.1939 0.3667 2.1939 1.4812
No log 1.9298 110 1.2678 0.5324 1.2678 1.1260
No log 1.9649 112 0.9733 0.6099 0.9733 0.9866
No log 2.0 114 0.9738 0.6099 0.9738 0.9868
No log 2.0351 116 0.9436 0.6056 0.9436 0.9714
No log 2.0702 118 1.0583 0.6579 1.0583 1.0287
No log 2.1053 120 1.3084 0.6310 1.3084 1.1439
No log 2.1404 122 1.6044 0.5140 1.6044 1.2666
No log 2.1754 124 1.5200 0.5227 1.5200 1.2329
No log 2.2105 126 1.1219 0.6543 1.1219 1.0592
No log 2.2456 128 0.9983 0.6447 0.9983 0.9991
No log 2.2807 130 1.0304 0.6829 1.0304 1.0151
No log 2.3158 132 0.9814 0.6784 0.9814 0.9906
No log 2.3509 134 1.1659 0.6374 1.1659 1.0798
No log 2.3860 136 1.2427 0.6374 1.2427 1.1148
No log 2.4211 138 1.5310 0.5758 1.5310 1.2373
No log 2.4561 140 1.4172 0.5907 1.4172 1.1905
No log 2.4912 142 0.8330 0.7333 0.8330 0.9127
No log 2.5263 144 0.6174 0.8144 0.6174 0.7857
No log 2.5614 146 0.6913 0.7561 0.6913 0.8314
No log 2.5965 148 0.8655 0.6709 0.8655 0.9303
No log 2.6316 150 0.8575 0.6709 0.8575 0.9260
No log 2.6667 152 0.8625 0.6709 0.8625 0.9287
No log 2.7018 154 0.9166 0.6829 0.9166 0.9574
No log 2.7368 156 0.7252 0.7044 0.7252 0.8516
No log 2.7719 158 0.5911 0.8054 0.5911 0.7688
No log 2.8070 160 0.6187 0.7871 0.6187 0.7866
No log 2.8421 162 0.6689 0.7515 0.6689 0.8179
No log 2.8772 164 0.8842 0.6905 0.8842 0.9403
No log 2.9123 166 1.4531 0.6105 1.4531 1.2054
No log 2.9474 168 1.3857 0.5967 1.3857 1.1772
No log 2.9825 170 0.8695 0.6905 0.8695 0.9325
No log 3.0175 172 0.6514 0.7922 0.6514 0.8071
No log 3.0526 174 0.6935 0.7651 0.6935 0.8328
No log 3.0877 176 0.6548 0.7895 0.6548 0.8092
No log 3.1228 178 0.6577 0.8199 0.6577 0.8110
No log 3.1579 180 0.9276 0.6705 0.9276 0.9631
No log 3.1930 182 0.9999 0.6705 0.9999 0.9999
No log 3.2281 184 0.7153 0.7889 0.7153 0.8458
No log 3.2632 186 0.6018 0.8272 0.6018 0.7758
No log 3.2982 188 0.6022 0.8171 0.6022 0.7760
No log 3.3333 190 0.7399 0.7701 0.7399 0.8602
No log 3.3684 192 1.0148 0.6592 1.0148 1.0074
No log 3.4035 194 1.0893 0.6441 1.0893 1.0437
No log 3.4386 196 0.8316 0.6901 0.8316 0.9119
No log 3.4737 198 0.6864 0.8068 0.6864 0.8285
No log 3.5088 200 0.6923 0.8068 0.6923 0.8320
No log 3.5439 202 0.6542 0.8182 0.6542 0.8088
No log 3.5789 204 0.7042 0.7910 0.7042 0.8391
No log 3.6140 206 0.6361 0.8372 0.6361 0.7975
No log 3.6491 208 0.5456 0.8434 0.5456 0.7387
No log 3.6842 210 0.5372 0.8395 0.5372 0.7329
No log 3.7193 212 0.5291 0.8395 0.5291 0.7274
No log 3.7544 214 0.5213 0.825 0.5213 0.7220
No log 3.7895 216 0.5755 0.8447 0.5755 0.7586
No log 3.8246 218 0.5645 0.8395 0.5645 0.7513
No log 3.8596 220 0.5492 0.8434 0.5492 0.7411
No log 3.8947 222 0.7384 0.7640 0.7384 0.8593
No log 3.9298 224 1.0672 0.6839 1.0672 1.0331
No log 3.9649 226 0.9332 0.7196 0.9332 0.9660
No log 4.0 228 0.8591 0.7273 0.8591 0.9269
No log 4.0351 230 0.5418 0.8343 0.5418 0.7360
No log 4.0702 232 0.5022 0.8409 0.5022 0.7086
No log 4.1053 234 0.5233 0.8506 0.5233 0.7234
No log 4.1404 236 0.5448 0.8655 0.5448 0.7381
No log 4.1754 238 0.6173 0.8488 0.6173 0.7857
No log 4.2105 240 0.7660 0.7262 0.7660 0.8752
No log 4.2456 242 0.8059 0.6864 0.8059 0.8977
No log 4.2807 244 0.7208 0.7711 0.7208 0.8490
No log 4.3158 246 0.7123 0.8105 0.7123 0.8440
No log 4.3509 248 0.7212 0.8105 0.7212 0.8492
No log 4.3860 250 0.7149 0.8588 0.7149 0.8455
No log 4.4211 252 0.7141 0.8521 0.7141 0.8451
No log 4.4561 254 0.6929 0.8521 0.6929 0.8324
No log 4.4912 256 0.6828 0.8343 0.6828 0.8263
No log 4.5263 258 0.8556 0.6705 0.8556 0.9250
No log 4.5614 260 0.9458 0.6590 0.9458 0.9725
No log 4.5965 262 0.7597 0.7283 0.7597 0.8716
No log 4.6316 264 0.5637 0.8205 0.5637 0.7508
No log 4.6667 266 0.6424 0.7448 0.6424 0.8015
No log 4.7018 268 0.6709 0.7172 0.6709 0.8191
No log 4.7368 270 0.6016 0.8 0.6016 0.7756
No log 4.7719 272 0.5680 0.8054 0.5680 0.7536
No log 4.8070 274 0.5417 0.8383 0.5417 0.7360
No log 4.8421 276 0.5100 0.8503 0.5100 0.7141
No log 4.8772 278 0.5262 0.8503 0.5262 0.7254
No log 4.9123 280 0.5428 0.8503 0.5428 0.7367
No log 4.9474 282 0.6354 0.8075 0.6354 0.7971
No log 4.9825 284 0.6968 0.7662 0.6968 0.8348
No log 5.0175 286 0.7249 0.7083 0.7249 0.8514
No log 5.0526 288 0.7885 0.6087 0.7885 0.8880
No log 5.0877 290 0.8370 0.6015 0.8370 0.9149
No log 5.1228 292 0.9504 0.4762 0.9504 0.9749
No log 5.1579 294 0.8739 0.6212 0.8739 0.9348
No log 5.1930 296 0.7844 0.6429 0.7844 0.8857
No log 5.2281 298 0.8329 0.6624 0.8329 0.9126
No log 5.2632 300 0.8610 0.6832 0.8610 0.9279
No log 5.2982 302 0.7122 0.7485 0.7122 0.8439
No log 5.3333 304 0.5887 0.8158 0.5887 0.7672
No log 5.3684 306 0.6791 0.7619 0.6791 0.8241
No log 5.4035 308 0.6857 0.7619 0.6857 0.8280
No log 5.4386 310 0.6162 0.7919 0.6162 0.7850
No log 5.4737 312 0.7000 0.7531 0.7000 0.8367
No log 5.5088 314 0.9094 0.6748 0.9094 0.9536
No log 5.5439 316 0.8973 0.6748 0.8973 0.9473
No log 5.5789 318 0.7712 0.7160 0.7712 0.8782
No log 5.6140 320 0.7020 0.7778 0.7020 0.8379
No log 5.6491 322 0.6630 0.7952 0.6630 0.8143
No log 5.6842 324 0.6365 0.8144 0.6365 0.7978
No log 5.7193 326 0.6365 0.7952 0.6365 0.7978
No log 5.7544 328 0.5883 0.8402 0.5883 0.7670
No log 5.7895 330 0.5609 0.8402 0.5609 0.7489
No log 5.8246 332 0.5694 0.8229 0.5694 0.7546
No log 5.8596 334 0.5880 0.8249 0.5880 0.7668
No log 5.8947 336 0.6154 0.8 0.6154 0.7845
No log 5.9298 338 0.5773 0.8421 0.5773 0.7598
No log 5.9649 340 0.5525 0.8205 0.5525 0.7433
No log 6.0 342 0.5867 0.8027 0.5867 0.7660
No log 6.0351 344 0.6139 0.7973 0.6139 0.7835
No log 6.0702 346 0.6971 0.75 0.6971 0.8350
No log 6.1053 348 0.7129 0.7784 0.7129 0.8444
No log 6.1404 350 0.6309 0.8023 0.6309 0.7943
No log 6.1754 352 0.5916 0.8457 0.5916 0.7691
No log 6.2105 354 0.5670 0.8221 0.5670 0.7530
No log 6.2456 356 0.5872 0.8148 0.5872 0.7663
No log 6.2807 358 0.6070 0.8276 0.6070 0.7791
No log 6.3158 360 0.6912 0.7907 0.6912 0.8314
No log 6.3509 362 0.7531 0.7356 0.7531 0.8678
No log 6.3860 364 0.6920 0.7647 0.6920 0.8319
No log 6.4211 366 0.6240 0.7871 0.6240 0.7899
No log 6.4561 368 0.6160 0.8133 0.6160 0.7849
No log 6.4912 370 0.6160 0.7919 0.6160 0.7848
No log 6.5263 372 0.5954 0.8133 0.5954 0.7716
No log 6.5614 374 0.6349 0.7651 0.6349 0.7968
No log 6.5965 376 0.6694 0.7516 0.6694 0.8182
No log 6.6316 378 0.8032 0.6988 0.8032 0.8962
No log 6.6667 380 0.8243 0.7066 0.8243 0.9079
No log 6.7018 382 0.7574 0.6826 0.7574 0.8703
No log 6.7368 384 0.6116 0.8047 0.6116 0.7821
No log 6.7719 386 0.5836 0.8193 0.5836 0.7639
No log 6.8070 388 0.5954 0.8129 0.5954 0.7716
No log 6.8421 390 0.6590 0.7898 0.6590 0.8118
No log 6.8772 392 0.8327 0.7073 0.8327 0.9125
No log 6.9123 394 0.9214 0.6706 0.9214 0.9599
No log 6.9474 396 0.8021 0.7117 0.8021 0.8956
No log 6.9825 398 0.6731 0.7448 0.6731 0.8204
No log 7.0175 400 0.7136 0.75 0.7136 0.8448
No log 7.0526 402 0.7219 0.7361 0.7219 0.8497
No log 7.0877 404 0.6496 0.7639 0.6496 0.8060
No log 7.1228 406 0.6388 0.7733 0.6388 0.7993
No log 7.1579 408 0.6974 0.7564 0.6974 0.8351
No log 7.1930 410 0.6970 0.7417 0.6970 0.8349
No log 7.2281 412 0.6617 0.7682 0.6617 0.8134
No log 7.2632 414 0.6129 0.7891 0.6129 0.7829
No log 7.2982 416 0.5827 0.7949 0.5827 0.7634
No log 7.3333 418 0.6346 0.7799 0.6346 0.7966
No log 7.3684 420 0.6783 0.7625 0.6783 0.8236
No log 7.4035 422 0.6770 0.7529 0.6770 0.8228
No log 7.4386 424 0.6140 0.7975 0.6140 0.7836
No log 7.4737 426 0.5929 0.7799 0.5929 0.7700
No log 7.5088 428 0.6063 0.7848 0.6063 0.7787
No log 7.5439 430 0.6756 0.7683 0.6756 0.8220
No log 7.5789 432 0.7069 0.7394 0.7069 0.8408
No log 7.6140 434 0.8196 0.7066 0.8196 0.9053
No log 7.6491 436 0.9147 0.6982 0.9147 0.9564
No log 7.6842 438 0.8516 0.7135 0.8516 0.9228
No log 7.7193 440 0.7218 0.7977 0.7218 0.8496
No log 7.7544 442 0.6670 0.8272 0.6670 0.8167
No log 7.7895 444 0.5990 0.8228 0.5990 0.7740
No log 7.8246 446 0.5902 0.8026 0.5902 0.7682
No log 7.8596 448 0.6110 0.7552 0.6110 0.7817
No log 7.8947 450 0.6405 0.7606 0.6405 0.8003
No log 7.9298 452 0.6613 0.7606 0.6613 0.8132
No log 7.9649 454 0.6505 0.7465 0.6505 0.8065
No log 8.0 456 0.6424 0.7413 0.6424 0.8015
No log 8.0351 458 0.6366 0.7619 0.6366 0.7979
No log 8.0702 460 0.5993 0.8026 0.5993 0.7742
No log 8.1053 462 0.5298 0.8205 0.5298 0.7279
No log 8.1404 464 0.4945 0.8302 0.4945 0.7032
No log 8.1754 466 0.4798 0.8221 0.4798 0.6927
No log 8.2105 468 0.4995 0.8171 0.4995 0.7068
No log 8.2456 470 0.5316 0.8372 0.5316 0.7291
No log 8.2807 472 0.5768 0.8304 0.5768 0.7594
No log 8.3158 474 0.5907 0.8214 0.5907 0.7686
No log 8.3509 476 0.5686 0.8228 0.5686 0.7540
No log 8.3860 478 0.5602 0.8 0.5602 0.7485
No log 8.4211 480 0.5882 0.8026 0.5882 0.7669
No log 8.4561 482 0.6450 0.7843 0.6450 0.8031
No log 8.4912 484 0.6569 0.7733 0.6569 0.8105
No log 8.5263 486 0.6799 0.7517 0.6799 0.8246
No log 8.5614 488 0.6902 0.7451 0.6902 0.8308
No log 8.5965 490 0.7097 0.75 0.7097 0.8424
No log 8.6316 492 0.6964 0.7561 0.6964 0.8345
No log 8.6667 494 0.6419 0.7595 0.6419 0.8012
No log 8.7018 496 0.5807 0.7974 0.5807 0.7621
No log 8.7368 498 0.5702 0.7974 0.5702 0.7551
0.4595 8.7719 500 0.5994 0.7799 0.5994 0.7742
0.4595 8.8070 502 0.6873 0.7578 0.6873 0.8290
0.4595 8.8421 504 0.7275 0.7375 0.7275 0.8529
0.4595 8.8772 506 0.6494 0.7403 0.6494 0.8059
0.4595 8.9123 508 0.5779 0.7838 0.5779 0.7602
0.4595 8.9474 510 0.5777 0.7724 0.5777 0.7601
0.4595 8.9825 512 0.5786 0.7917 0.5786 0.7607
0.4595 9.0175 514 0.6316 0.7517 0.6316 0.7947
0.4595 9.0526 516 0.8081 0.6014 0.8081 0.8990
0.4595 9.0877 518 0.9140 0.64 0.9140 0.9560
0.4595 9.1228 520 0.9078 0.64 0.9078 0.9528
0.4595 9.1579 522 0.8079 0.6301 0.8079 0.8988

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k12_task1_organization

Finetuned
(4019)
this model