ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k3_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8382
  • Qwk: 0.7105
  • Mse: 0.8382
  • Rmse: 0.9155

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 6.8079 0.0232 6.8079 2.6092
No log 0.3077 4 3.9957 0.0721 3.9957 1.9989
No log 0.4615 6 2.4964 0.1467 2.4964 1.5800
No log 0.6154 8 1.9539 0.1613 1.9539 1.3978
No log 0.7692 10 1.6517 0.1887 1.6517 1.2852
No log 0.9231 12 1.6556 0.1887 1.6556 1.2867
No log 1.0769 14 1.8431 0.3115 1.8431 1.3576
No log 1.2308 16 2.4402 0.1224 2.4402 1.5621
No log 1.3846 18 2.3083 0.0959 2.3083 1.5193
No log 1.5385 20 2.0433 0.2836 2.0433 1.4295
No log 1.6923 22 1.8370 0.2969 1.8370 1.3554
No log 1.8462 24 1.5177 0.3276 1.5177 1.2320
No log 2.0 26 1.3497 0.3363 1.3497 1.1618
No log 2.1538 28 1.3575 0.3333 1.3575 1.1651
No log 2.3077 30 1.4291 0.3697 1.4291 1.1954
No log 2.4615 32 1.5942 0.4394 1.5942 1.2626
No log 2.6154 34 1.7575 0.3881 1.7575 1.3257
No log 2.7692 36 1.8834 0.3803 1.8834 1.3724
No log 2.9231 38 2.1526 0.2963 2.1526 1.4672
No log 3.0769 40 1.9671 0.3893 1.9671 1.4025
No log 3.2308 42 1.6066 0.4394 1.6066 1.2675
No log 3.3846 44 1.4053 0.3051 1.4053 1.1855
No log 3.5385 46 1.3314 0.2957 1.3314 1.1539
No log 3.6923 48 1.4127 0.3902 1.4127 1.1886
No log 3.8462 50 1.5551 0.4252 1.5551 1.2470
No log 4.0 52 1.8076 0.3433 1.8076 1.3445
No log 4.1538 54 1.7420 0.3333 1.7420 1.3198
No log 4.3077 56 1.3761 0.4444 1.3761 1.1731
No log 4.4615 58 1.2353 0.4480 1.2353 1.1114
No log 4.6154 60 1.2185 0.4516 1.2185 1.1039
No log 4.7692 62 1.2712 0.4724 1.2712 1.1275
No log 4.9231 64 1.4150 0.5676 1.4150 1.1895
No log 5.0769 66 1.2657 0.6 1.2657 1.1250
No log 5.2308 68 1.0764 0.5882 1.0764 1.0375
No log 5.3846 70 1.0543 0.6122 1.0543 1.0268
No log 5.5385 72 1.0472 0.5906 1.0472 1.0233
No log 5.6923 74 1.0709 0.5874 1.0709 1.0348
No log 5.8462 76 1.1477 0.5694 1.1477 1.0713
No log 6.0 78 1.4537 0.4868 1.4537 1.2057
No log 6.1538 80 1.2745 0.5442 1.2745 1.1289
No log 6.3077 82 1.0920 0.5839 1.0920 1.0450
No log 6.4615 84 1.1372 0.6014 1.1372 1.0664
No log 6.6154 86 1.1154 0.6131 1.1154 1.0561
No log 6.7692 88 1.1225 0.6119 1.1225 1.0595
No log 6.9231 90 1.0869 0.6309 1.0869 1.0426
No log 7.0769 92 1.3275 0.5342 1.3275 1.1522
No log 7.2308 94 1.5130 0.5570 1.5130 1.2300
No log 7.3846 96 1.0637 0.6 1.0637 1.0314
No log 7.5385 98 0.9187 0.6331 0.9187 0.9585
No log 7.6923 100 0.9417 0.5970 0.9417 0.9704
No log 7.8462 102 1.0401 0.5909 1.0401 1.0199
No log 8.0 104 1.3743 0.4658 1.3743 1.1723
No log 8.1538 106 1.6433 0.4371 1.6433 1.2819
No log 8.3077 108 1.5257 0.4698 1.5257 1.2352
No log 8.4615 110 1.1590 0.5594 1.1590 1.0766
No log 8.6154 112 0.9689 0.6069 0.9689 0.9843
No log 8.7692 114 0.9401 0.7006 0.9401 0.9696
No log 8.9231 116 1.0150 0.6512 1.0150 1.0075
No log 9.0769 118 0.9807 0.6705 0.9807 0.9903
No log 9.2308 120 0.9296 0.6667 0.9296 0.9642
No log 9.3846 122 0.9442 0.6709 0.9442 0.9717
No log 9.5385 124 0.9441 0.6752 0.9441 0.9717
No log 9.6923 126 0.9060 0.6624 0.9060 0.9518
No log 9.8462 128 0.9541 0.6452 0.9541 0.9768
No log 10.0 130 1.0246 0.6358 1.0246 1.0122
No log 10.1538 132 1.0037 0.6795 1.0037 1.0018
No log 10.3077 134 0.9379 0.7081 0.9379 0.9685
No log 10.4615 136 0.9398 0.6951 0.9398 0.9694
No log 10.6154 138 0.8940 0.7186 0.8940 0.9455
No log 10.7692 140 0.9106 0.6923 0.9106 0.9542
No log 10.9231 142 0.9746 0.7174 0.9746 0.9872
No log 11.0769 144 1.0429 0.7174 1.0429 1.0212
No log 11.2308 146 1.0552 0.6705 1.0552 1.0272
No log 11.3846 148 0.9650 0.6786 0.9650 0.9823
No log 11.5385 150 0.9038 0.7101 0.9038 0.9507
No log 11.6923 152 0.8793 0.6832 0.8793 0.9377
No log 11.8462 154 0.8763 0.7296 0.8763 0.9361
No log 12.0 156 0.9549 0.6982 0.9549 0.9772
No log 12.1538 158 0.9524 0.7125 0.9524 0.9759
No log 12.3077 160 0.8976 0.7134 0.8976 0.9474
No log 12.4615 162 0.9311 0.7134 0.9311 0.9649
No log 12.6154 164 1.0665 0.6667 1.0665 1.0327
No log 12.7692 166 1.0685 0.6705 1.0685 1.0337
No log 12.9231 168 1.0215 0.7125 1.0215 1.0107
No log 13.0769 170 1.0938 0.6027 1.0938 1.0459
No log 13.2308 172 1.1917 0.4493 1.1917 1.0917
No log 13.3846 174 1.1656 0.5578 1.1656 1.0796
No log 13.5385 176 1.0819 0.6839 1.0819 1.0401
No log 13.6923 178 0.9531 0.6753 0.9531 0.9763
No log 13.8462 180 0.8914 0.6483 0.8914 0.9441
No log 14.0 182 0.8724 0.6897 0.8724 0.9340
No log 14.1538 184 0.8508 0.6667 0.8508 0.9224
No log 14.3077 186 0.8421 0.6531 0.8421 0.9177
No log 14.4615 188 1.0115 0.6667 1.0115 1.0057
No log 14.6154 190 1.2071 0.6136 1.2071 1.0987
No log 14.7692 192 1.0596 0.6889 1.0596 1.0294
No log 14.9231 194 0.8952 0.7089 0.8952 0.9462
No log 15.0769 196 0.9428 0.6154 0.9428 0.9710
No log 15.2308 198 0.9839 0.5778 0.9839 0.9919
No log 15.3846 200 1.0311 0.6383 1.0311 1.0154
No log 15.5385 202 1.0906 0.6081 1.0906 1.0443
No log 15.6923 204 1.0684 0.6225 1.0684 1.0336
No log 15.8462 206 0.9898 0.6447 0.9898 0.9949
No log 16.0 208 0.9690 0.6301 0.9690 0.9844
No log 16.1538 210 0.9611 0.6531 0.9611 0.9804
No log 16.3077 212 0.9403 0.6301 0.9403 0.9697
No log 16.4615 214 0.9497 0.6755 0.9497 0.9745
No log 16.6154 216 0.9631 0.7006 0.9631 0.9814
No log 16.7692 218 0.9746 0.7160 0.9746 0.9872
No log 16.9231 220 0.9635 0.6795 0.9635 0.9816
No log 17.0769 222 0.9179 0.6623 0.9179 0.9581
No log 17.2308 224 0.9124 0.6623 0.9124 0.9552
No log 17.3846 226 0.9419 0.6795 0.9419 0.9705
No log 17.5385 228 0.9713 0.7117 0.9713 0.9855
No log 17.6923 230 0.9809 0.6795 0.9809 0.9904
No log 17.8462 232 0.9501 0.6839 0.9501 0.9747
No log 18.0 234 0.8899 0.6623 0.8899 0.9434
No log 18.1538 236 0.8716 0.6623 0.8716 0.9336
No log 18.3077 238 0.8867 0.6839 0.8867 0.9416
No log 18.4615 240 0.9222 0.6795 0.9222 0.9603
No log 18.6154 242 0.9412 0.7081 0.9412 0.9702
No log 18.7692 244 0.8941 0.6835 0.8941 0.9456
No log 18.9231 246 0.8661 0.6623 0.8661 0.9307
No log 19.0769 248 0.8817 0.6623 0.8817 0.9390
No log 19.2308 250 0.9447 0.6536 0.9447 0.9719
No log 19.3846 252 0.9659 0.6351 0.9659 0.9828
No log 19.5385 254 1.0210 0.6405 1.0210 1.0104
No log 19.6923 256 1.0271 0.6710 1.0271 1.0135
No log 19.8462 258 0.9689 0.7013 0.9689 0.9844
No log 20.0 260 0.9000 0.6667 0.9000 0.9487
No log 20.1538 262 0.8388 0.6528 0.8388 0.9159
No log 20.3077 264 0.8077 0.6986 0.8077 0.8987
No log 20.4615 266 0.8065 0.7308 0.8065 0.8981
No log 20.6154 268 0.9503 0.7081 0.9503 0.9748
No log 20.7692 270 1.1697 0.5747 1.1697 1.0815
No log 20.9231 272 1.1385 0.6118 1.1385 1.0670
No log 21.0769 274 0.9519 0.7089 0.9519 0.9756
No log 21.2308 276 0.8494 0.6839 0.8494 0.9216
No log 21.3846 278 0.8240 0.7097 0.8240 0.9078
No log 21.5385 280 0.8794 0.6835 0.8794 0.9378
No log 21.6923 282 1.0186 0.6667 1.0186 1.0093
No log 21.8462 284 1.0392 0.6448 1.0392 1.0194
No log 22.0 286 0.9409 0.7108 0.9409 0.9700
No log 22.1538 288 0.8791 0.6980 0.8791 0.9376
No log 22.3077 290 0.8717 0.6620 0.8717 0.9337
No log 22.4615 292 0.8907 0.6119 0.8907 0.9438
No log 22.6154 294 0.8940 0.6331 0.8940 0.9455
No log 22.7692 296 0.9241 0.6301 0.9241 0.9613
No log 22.9231 298 1.0230 0.6497 1.0230 1.0114
No log 23.0769 300 1.0990 0.6592 1.0990 1.0483
No log 23.2308 302 1.0633 0.6557 1.0633 1.0312
No log 23.3846 304 0.9251 0.7111 0.9251 0.9618
No log 23.5385 306 0.8355 0.7284 0.8355 0.9141
No log 23.6923 308 0.8409 0.7237 0.8409 0.9170
No log 23.8462 310 0.8617 0.6846 0.8617 0.9283
No log 24.0 312 0.9012 0.6241 0.9012 0.9493
No log 24.1538 314 0.9396 0.6383 0.9396 0.9693
No log 24.3077 316 0.9725 0.6345 0.9725 0.9861
No log 24.4615 318 1.0131 0.6301 1.0131 1.0065
No log 24.6154 320 1.0337 0.6753 1.0337 1.0167
No log 24.7692 322 0.9863 0.6624 0.9863 0.9931
No log 24.9231 324 0.9043 0.6846 0.9043 0.9510
No log 25.0769 326 0.8595 0.6803 0.8595 0.9271
No log 25.2308 328 0.8701 0.6622 0.8701 0.9328
No log 25.3846 330 0.9139 0.6483 0.9139 0.9560
No log 25.5385 332 1.0165 0.6301 1.0165 1.0082
No log 25.6923 334 1.0833 0.6038 1.0833 1.0408
No log 25.8462 336 1.0599 0.6145 1.0599 1.0295
No log 26.0 338 0.9776 0.6424 0.9776 0.9887
No log 26.1538 340 0.9451 0.6667 0.9451 0.9721
No log 26.3077 342 0.9488 0.6923 0.9488 0.9741
No log 26.4615 344 0.9290 0.7013 0.9290 0.9639
No log 26.6154 346 0.9140 0.6806 0.9140 0.9560
No log 26.7692 348 0.8986 0.6423 0.8986 0.9479
No log 26.9231 350 0.9070 0.6232 0.9070 0.9523
No log 27.0769 352 0.8983 0.6809 0.8983 0.9478
No log 27.2308 354 0.8774 0.7075 0.8774 0.9367
No log 27.3846 356 0.8917 0.7075 0.8917 0.9443
No log 27.5385 358 0.9333 0.6707 0.9333 0.9661
No log 27.6923 360 1.0259 0.6554 1.0259 1.0129
No log 27.8462 362 1.0261 0.6556 1.0261 1.0129
No log 28.0 364 0.9815 0.7006 0.9815 0.9907
No log 28.1538 366 0.8998 0.7024 0.8998 0.9486
No log 28.3077 368 0.8670 0.6667 0.8670 0.9311
No log 28.4615 370 0.8279 0.7020 0.8279 0.9099
No log 28.6154 372 0.8330 0.7034 0.8330 0.9127
No log 28.7692 374 0.8555 0.6713 0.8555 0.9249
No log 28.9231 376 0.9239 0.6533 0.9239 0.9612
No log 29.0769 378 0.9642 0.6753 0.9642 0.9819
No log 29.2308 380 0.9368 0.6710 0.9368 0.9679
No log 29.3846 382 0.9408 0.6918 0.9408 0.9700
No log 29.5385 384 0.9124 0.6709 0.9124 0.9552
No log 29.6923 386 0.8706 0.6842 0.8706 0.9330
No log 29.8462 388 0.8449 0.7067 0.8449 0.9192
No log 30.0 390 0.8413 0.7059 0.8413 0.9172
No log 30.1538 392 0.8520 0.7013 0.8520 0.9230
No log 30.3077 394 0.8796 0.6883 0.8796 0.9379
No log 30.4615 396 0.9291 0.6962 0.9291 0.9639
No log 30.6154 398 0.9610 0.6839 0.9610 0.9803
No log 30.7692 400 0.9640 0.6667 0.9640 0.9819
No log 30.9231 402 0.9949 0.6577 0.9949 0.9974
No log 31.0769 404 0.9938 0.6259 0.9938 0.9969
No log 31.2308 406 0.9704 0.6358 0.9704 0.9851
No log 31.3846 408 0.9589 0.6497 0.9589 0.9792
No log 31.5385 410 0.9227 0.6289 0.9227 0.9606
No log 31.6923 412 0.9149 0.6364 0.9149 0.9565
No log 31.8462 414 0.9462 0.6144 0.9462 0.9727
No log 32.0 416 0.9977 0.6581 0.9977 0.9989
No log 32.1538 418 1.0295 0.6282 1.0295 1.0146
No log 32.3077 420 1.0215 0.6242 1.0215 1.0107
No log 32.4615 422 1.0006 0.6395 1.0006 1.0003
No log 32.6154 424 0.9725 0.5985 0.9725 0.9862
No log 32.7692 426 0.9531 0.6571 0.9531 0.9763
No log 32.9231 428 0.9503 0.6475 0.9503 0.9748
No log 33.0769 430 0.9678 0.5985 0.9678 0.9838
No log 33.2308 432 0.9531 0.6143 0.9531 0.9763
No log 33.3846 434 0.9631 0.6241 0.9631 0.9814
No log 33.5385 436 0.9936 0.6351 0.9936 0.9968
No log 33.6923 438 0.9877 0.6533 0.9877 0.9938
No log 33.8462 440 0.9171 0.6447 0.9171 0.9577
No log 34.0 442 0.8666 0.7051 0.8666 0.9309
No log 34.1538 444 0.8584 0.6875 0.8584 0.9265
No log 34.3077 446 0.8823 0.7326 0.8823 0.9393
No log 34.4615 448 0.9252 0.7006 0.9252 0.9619
No log 34.6154 450 0.9252 0.6901 0.9252 0.9619
No log 34.7692 452 0.8972 0.6875 0.8972 0.9472
No log 34.9231 454 0.8538 0.7143 0.8538 0.9240
No log 35.0769 456 0.8595 0.7143 0.8595 0.9271
No log 35.2308 458 0.8700 0.6842 0.8700 0.9327
No log 35.3846 460 0.8776 0.6533 0.8776 0.9368
No log 35.5385 462 0.8651 0.6755 0.8651 0.9301
No log 35.6923 464 0.8862 0.6711 0.8862 0.9414
No log 35.8462 466 0.8927 0.6753 0.8927 0.9449
No log 36.0 468 0.8713 0.6928 0.8713 0.9335
No log 36.1538 470 0.8743 0.6928 0.8743 0.9351
No log 36.3077 472 0.8952 0.6797 0.8952 0.9462
No log 36.4615 474 0.9147 0.6447 0.9147 0.9564
No log 36.6154 476 0.9025 0.6786 0.9025 0.9500
No log 36.7692 478 0.9104 0.6971 0.9104 0.9542
No log 36.9231 480 0.9040 0.7039 0.9040 0.9508
No log 37.0769 482 0.9196 0.7104 0.9196 0.9590
No log 37.2308 484 0.8962 0.7104 0.8962 0.9467
No log 37.3846 486 0.8201 0.7486 0.8201 0.9056
No log 37.5385 488 0.7601 0.7578 0.7601 0.8718
No log 37.6923 490 0.7607 0.7516 0.7607 0.8722
No log 37.8462 492 0.8077 0.7516 0.8077 0.8987
No log 38.0 494 0.9347 0.6795 0.9347 0.9668
No log 38.1538 496 1.0103 0.6405 1.0103 1.0051
No log 38.3077 498 0.9734 0.6494 0.9734 0.9866
0.2949 38.4615 500 0.8558 0.7190 0.8558 0.9251
0.2949 38.6154 502 0.7698 0.7550 0.7698 0.8774
0.2949 38.7692 504 0.7766 0.7123 0.7766 0.8813
0.2949 38.9231 506 0.7961 0.7027 0.7961 0.8922
0.2949 39.0769 508 0.8036 0.7027 0.8036 0.8965
0.2949 39.2308 510 0.8291 0.7451 0.8291 0.9105
0.2949 39.3846 512 0.8422 0.7226 0.8422 0.9177
0.2949 39.5385 514 0.8735 0.6918 0.8735 0.9346
0.2949 39.6923 516 0.8999 0.72 0.8999 0.9486
0.2949 39.8462 518 0.9012 0.7219 0.9012 0.9493
0.2949 40.0 520 0.8829 0.6962 0.8829 0.9396
0.2949 40.1538 522 0.8347 0.7097 0.8347 0.9136
0.2949 40.3077 524 0.8081 0.7451 0.8081 0.8989
0.2949 40.4615 526 0.8060 0.7451 0.8060 0.8978
0.2949 40.6154 528 0.8355 0.7020 0.8355 0.9140
0.2949 40.7692 530 0.8831 0.7226 0.8831 0.9398
0.2949 40.9231 532 0.8854 0.7226 0.8854 0.9409
0.2949 41.0769 534 0.8356 0.7237 0.8356 0.9141
0.2949 41.2308 536 0.7593 0.7347 0.7593 0.8714
0.2949 41.3846 538 0.7127 0.7222 0.7127 0.8442
0.2949 41.5385 540 0.6928 0.7467 0.6928 0.8324
0.2949 41.6923 542 0.6988 0.7451 0.6988 0.8359
0.2949 41.8462 544 0.7363 0.7484 0.7363 0.8581
0.2949 42.0 546 0.7911 0.7531 0.7911 0.8895
0.2949 42.1538 548 0.8237 0.7607 0.8237 0.9076
0.2949 42.3077 550 0.8169 0.7643 0.8169 0.9038
0.2949 42.4615 552 0.8033 0.7355 0.8033 0.8963
0.2949 42.6154 554 0.7763 0.7190 0.7763 0.8811
0.2949 42.7692 556 0.7696 0.7237 0.7696 0.8773
0.2949 42.9231 558 0.7625 0.7451 0.7625 0.8732
0.2949 43.0769 560 0.7652 0.7451 0.7652 0.8748
0.2949 43.2308 562 0.7962 0.7237 0.7962 0.8923
0.2949 43.3846 564 0.8269 0.7190 0.8269 0.9093
0.2949 43.5385 566 0.8426 0.7051 0.8426 0.9179
0.2949 43.6923 568 0.8452 0.7051 0.8452 0.9193
0.2949 43.8462 570 0.8360 0.7152 0.8360 0.9144
0.2949 44.0 572 0.8382 0.7105 0.8382 0.9155

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k3_task1_organization

Finetuned
(4023)
this model