ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k17_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8347
  • Qwk: 0.725
  • Mse: 0.8347
  • Rmse: 0.9136

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.025 2 6.6183 0.0308 6.6183 2.5726
No log 0.05 4 4.0742 0.0779 4.0742 2.0185
No log 0.075 6 2.7883 0.0633 2.7883 1.6698
No log 0.1 8 2.1851 0.1053 2.1851 1.4782
No log 0.125 10 1.9752 0.2373 1.9752 1.4054
No log 0.15 12 1.7519 0.2162 1.7519 1.3236
No log 0.175 14 1.7923 0.1698 1.7923 1.3388
No log 0.2 16 1.7458 0.2586 1.7458 1.3213
No log 0.225 18 2.7517 0.0774 2.7517 1.6588
No log 0.25 20 2.9969 0.0617 2.9969 1.7311
No log 0.275 22 2.3367 0.0704 2.3367 1.5286
No log 0.3 24 1.7617 0.3009 1.7617 1.3273
No log 0.325 26 1.7766 0.2364 1.7766 1.3329
No log 0.35 28 1.8635 0.2783 1.8635 1.3651
No log 0.375 30 1.8970 0.3279 1.8970 1.3773
No log 0.4 32 1.9332 0.3175 1.9332 1.3904
No log 0.425 34 1.8191 0.3307 1.8191 1.3487
No log 0.45 36 1.8123 0.3307 1.8123 1.3462
No log 0.475 38 1.5925 0.3360 1.5925 1.2620
No log 0.5 40 2.0246 0.3425 2.0246 1.4229
No log 0.525 42 2.1265 0.2933 2.1265 1.4582
No log 0.55 44 1.9712 0.3537 1.9712 1.4040
No log 0.575 46 1.3645 0.4526 1.3645 1.1681
No log 0.6 48 1.2598 0.5263 1.2598 1.1224
No log 0.625 50 1.7295 0.4247 1.7295 1.3151
No log 0.65 52 2.6510 0.2813 2.6510 1.6282
No log 0.675 54 2.5977 0.2974 2.5977 1.6118
No log 0.7 56 2.4413 0.2769 2.4413 1.5625
No log 0.725 58 2.2990 0.2903 2.2990 1.5163
No log 0.75 60 1.7304 0.4348 1.7304 1.3154
No log 0.775 62 1.3075 0.4706 1.3075 1.1435
No log 0.8 64 1.2333 0.4923 1.2333 1.1105
No log 0.825 66 1.2342 0.4762 1.2342 1.1109
No log 0.85 68 1.1440 0.4697 1.1440 1.0696
No log 0.875 70 1.2329 0.5306 1.2329 1.1104
No log 0.9 72 1.4216 0.5584 1.4216 1.1923
No log 0.925 74 1.4000 0.5466 1.4000 1.1832
No log 0.95 76 1.2719 0.5696 1.2719 1.1278
No log 0.975 78 1.5131 0.5422 1.5131 1.2301
No log 1.0 80 1.6490 0.4908 1.6490 1.2841
No log 1.025 82 1.5338 0.5641 1.5338 1.2385
No log 1.05 84 1.1320 0.6081 1.1320 1.0639
No log 1.075 86 0.9126 0.6577 0.9126 0.9553
No log 1.1 88 0.8448 0.7067 0.8448 0.9191
No log 1.125 90 0.8732 0.6351 0.8732 0.9345
No log 1.15 92 1.2087 0.5986 1.2087 1.0994
No log 1.175 94 1.3980 0.5676 1.3980 1.1824
No log 1.2 96 1.1995 0.5946 1.1995 1.0952
No log 1.225 98 1.0039 0.64 1.0039 1.0020
No log 1.25 100 0.9099 0.6577 0.9099 0.9539
No log 1.275 102 0.8930 0.64 0.8930 0.9450
No log 1.3 104 0.9772 0.6623 0.9772 0.9885
No log 1.325 106 1.1492 0.6234 1.1492 1.0720
No log 1.35 108 1.4311 0.5902 1.4311 1.1963
No log 1.375 110 1.1565 0.6441 1.1565 1.0754
No log 1.4 112 0.8354 0.7326 0.8354 0.9140
No log 1.425 114 0.7980 0.7399 0.7980 0.8933
No log 1.45 116 1.0652 0.6703 1.0652 1.0321
No log 1.475 118 1.1177 0.6595 1.1177 1.0572
No log 1.5 120 0.8234 0.7545 0.8234 0.9074
No log 1.525 122 0.7576 0.7172 0.7576 0.8704
No log 1.55 124 0.9182 0.6389 0.9182 0.9582
No log 1.575 126 0.8437 0.7075 0.8437 0.9185
No log 1.6 128 1.0414 0.6027 1.0414 1.0205
No log 1.625 130 1.4435 0.5389 1.4435 1.2015
No log 1.65 132 1.3833 0.5542 1.3833 1.1762
No log 1.675 134 1.2556 0.6145 1.2556 1.1205
No log 1.7 136 0.9171 0.6835 0.9171 0.9576
No log 1.725 138 0.7882 0.7237 0.7882 0.8878
No log 1.75 140 0.7456 0.7712 0.7456 0.8635
No log 1.775 142 0.7000 0.7712 0.7000 0.8367
No log 1.8 144 0.6810 0.7871 0.6810 0.8252
No log 1.825 146 0.6610 0.7898 0.6610 0.8130
No log 1.85 148 0.6723 0.8121 0.6723 0.8199
No log 1.875 150 0.6658 0.7898 0.6658 0.8159
No log 1.9 152 0.7711 0.6933 0.7711 0.8781
No log 1.925 154 1.0443 0.5556 1.0443 1.0219
No log 1.95 156 1.0527 0.5556 1.0527 1.0260
No log 1.975 158 0.8372 0.6853 0.8372 0.9150
No log 2.0 160 0.8034 0.7248 0.8034 0.8963
No log 2.025 162 0.9878 0.6364 0.9878 0.9939
No log 2.05 164 0.9834 0.6410 0.9834 0.9916
No log 2.075 166 0.8695 0.6759 0.8695 0.9324
No log 2.1 168 0.8082 0.7123 0.8082 0.8990
No log 2.125 170 0.8071 0.7114 0.8071 0.8984
No log 2.15 172 0.9461 0.7160 0.9461 0.9727
No log 2.175 174 1.3464 0.5665 1.3464 1.1604
No log 2.2 176 1.6261 0.5424 1.6261 1.2752
No log 2.225 178 1.2834 0.5862 1.2834 1.1329
No log 2.25 180 0.8094 0.7394 0.8094 0.8997
No log 2.275 182 0.6705 0.7692 0.6705 0.8189
No log 2.3 184 0.6274 0.7898 0.6274 0.7921
No log 2.325 186 0.6340 0.8 0.6340 0.7962
No log 2.35 188 0.7165 0.7556 0.7165 0.8465
No log 2.375 190 0.9934 0.7403 0.9934 0.9967
No log 2.4 192 1.1444 0.7 1.1444 1.0698
No log 2.425 194 1.3219 0.6328 1.3219 1.1498
No log 2.45 196 1.1478 0.6667 1.1478 1.0714
No log 2.475 198 1.0438 0.6782 1.0438 1.0216
No log 2.5 200 0.9596 0.6860 0.9596 0.9796
No log 2.525 202 1.0474 0.6893 1.0474 1.0234
No log 2.55 204 0.9735 0.7079 0.9735 0.9867
No log 2.575 206 0.9034 0.7151 0.9034 0.9505
No log 2.6 208 0.7757 0.7711 0.7757 0.8807
No log 2.625 210 0.6876 0.7517 0.6876 0.8292
No log 2.65 212 0.6889 0.7448 0.6889 0.8300
No log 2.675 214 0.7218 0.7517 0.7218 0.8496
No log 2.7 216 0.9010 0.6842 0.9010 0.9492
No log 2.725 218 0.9240 0.7117 0.9240 0.9613
No log 2.75 220 0.8583 0.7456 0.8583 0.9264
No log 2.775 222 0.9165 0.7065 0.9165 0.9574
No log 2.8 224 1.1170 0.6771 1.1170 1.0569
No log 2.825 226 1.1437 0.6597 1.1437 1.0695
No log 2.85 228 0.8195 0.7442 0.8195 0.9053
No log 2.875 230 0.6890 0.7927 0.6890 0.8300
No log 2.9 232 0.6602 0.7771 0.6602 0.8126
No log 2.925 234 0.7190 0.7643 0.7190 0.8479
No log 2.95 236 0.7409 0.7417 0.7409 0.8607
No log 2.975 238 0.7381 0.7383 0.7381 0.8591
No log 3.0 240 0.7218 0.7347 0.7218 0.8496
No log 3.025 242 0.7856 0.7042 0.7856 0.8863
No log 3.05 244 0.9087 0.6331 0.9087 0.9532
No log 3.075 246 0.7692 0.7 0.7692 0.8770
No log 3.1 248 0.7711 0.6716 0.7711 0.8781
No log 3.125 250 0.8285 0.6714 0.8285 0.9102
No log 3.15 252 0.8289 0.6713 0.8289 0.9105
No log 3.175 254 0.7741 0.7483 0.7741 0.8798
No log 3.2 256 0.7541 0.7821 0.7541 0.8684
No log 3.225 258 0.7470 0.7821 0.7470 0.8643
No log 3.25 260 0.7269 0.8025 0.7269 0.8526
No log 3.275 262 0.7115 0.7875 0.7115 0.8435
No log 3.3 264 0.6641 0.7733 0.6641 0.8149
No log 3.325 266 0.6844 0.7517 0.6844 0.8273
No log 3.35 268 0.7144 0.7771 0.7144 0.8452
No log 3.375 270 0.8375 0.7468 0.8375 0.9152
No log 3.4 272 1.0564 0.5696 1.0564 1.0278
No log 3.425 274 1.1741 0.5912 1.1741 1.0836
No log 3.45 276 1.0570 0.5419 1.0570 1.0281
No log 3.475 278 0.8242 0.7261 0.8242 0.9079
No log 3.5 280 0.7131 0.7671 0.7131 0.8445
No log 3.525 282 0.8135 0.6853 0.8135 0.9019
No log 3.55 284 0.8371 0.6713 0.8371 0.9150
No log 3.575 286 0.7933 0.7483 0.7933 0.8907
No log 3.6 288 0.8037 0.7516 0.8037 0.8965
No log 3.625 290 0.8887 0.6918 0.8887 0.9427
No log 3.65 292 0.8919 0.6988 0.8919 0.9444
No log 3.675 294 0.8433 0.7305 0.8433 0.9183
No log 3.7 296 0.7741 0.7545 0.7741 0.8798
No log 3.725 298 0.7518 0.7901 0.7518 0.8671
No log 3.75 300 0.7664 0.7826 0.7664 0.8754
No log 3.775 302 0.8133 0.7362 0.8133 0.9018
No log 3.8 304 1.0186 0.6705 1.0186 1.0093
No log 3.825 306 1.3485 0.6364 1.3485 1.1612
No log 3.85 308 1.3265 0.6354 1.3265 1.1517
No log 3.875 310 1.2561 0.6383 1.2561 1.1208
No log 3.9 312 1.0493 0.6593 1.0493 1.0244
No log 3.925 314 0.8931 0.7079 0.8931 0.9450
No log 3.95 316 0.8629 0.7108 0.8629 0.9289
No log 3.975 318 0.7999 0.7205 0.7999 0.8944
No log 4.0 320 0.7421 0.7547 0.7421 0.8614
No log 4.025 322 0.7274 0.7848 0.7274 0.8529
No log 4.05 324 0.7527 0.76 0.7527 0.8676
No log 4.075 326 0.7979 0.7310 0.7979 0.8932
No log 4.1 328 0.8259 0.7234 0.8259 0.9088
No log 4.125 330 0.8037 0.7273 0.8037 0.8965
No log 4.15 332 0.7864 0.6803 0.7864 0.8868
No log 4.175 334 0.7999 0.7179 0.7999 0.8944
No log 4.2 336 0.8315 0.7051 0.8315 0.9119
No log 4.225 338 0.8080 0.6846 0.8080 0.8989
No log 4.25 340 0.7879 0.7285 0.7879 0.8876
No log 4.275 342 0.8009 0.6933 0.8009 0.8950
No log 4.3 344 0.8137 0.6974 0.8137 0.9021
No log 4.325 346 0.8113 0.7089 0.8113 0.9007
No log 4.35 348 0.7506 0.7636 0.7506 0.8664
No log 4.375 350 0.7092 0.7976 0.7092 0.8421
No log 4.4 352 0.6805 0.8047 0.6805 0.8249
No log 4.425 354 0.6784 0.8121 0.6784 0.8237
No log 4.45 356 0.7618 0.7925 0.7618 0.8728
No log 4.475 358 0.7463 0.7763 0.7463 0.8639
No log 4.5 360 0.7026 0.7771 0.7026 0.8382
No log 4.525 362 0.7335 0.7636 0.7335 0.8564
No log 4.55 364 0.8769 0.6829 0.8769 0.9364
No log 4.575 366 0.9505 0.6909 0.9505 0.9749
No log 4.6 368 0.8804 0.7381 0.8804 0.9383
No log 4.625 370 0.7880 0.7456 0.7880 0.8877
No log 4.65 372 0.8318 0.7456 0.8318 0.9120
No log 4.675 374 0.7126 0.7647 0.7126 0.8441
No log 4.7 376 0.6198 0.7976 0.6198 0.7873
No log 4.725 378 0.5848 0.8070 0.5848 0.7647
No log 4.75 380 0.5745 0.7904 0.5745 0.7580
No log 4.775 382 0.6156 0.8193 0.6156 0.7846
No log 4.8 384 0.7368 0.7239 0.7368 0.8583
No log 4.825 386 0.9072 0.6928 0.9072 0.9525
No log 4.85 388 1.0326 0.6111 1.0326 1.0162
No log 4.875 390 1.0881 0.5816 1.0881 1.0431
No log 4.9 392 0.9761 0.6277 0.9761 0.9880
No log 4.925 394 0.8115 0.6806 0.8115 0.9008
No log 4.95 396 0.6928 0.7484 0.6928 0.8323
No log 4.975 398 0.8240 0.7195 0.8240 0.9078
No log 5.0 400 1.1796 0.6207 1.1796 1.0861
No log 5.025 402 1.2661 0.6127 1.2661 1.1252
No log 5.05 404 1.0918 0.6061 1.0918 1.0449
No log 5.075 406 0.8964 0.7030 0.8964 0.9468
No log 5.1 408 0.8476 0.7305 0.8476 0.9206
No log 5.125 410 0.8243 0.7108 0.8243 0.9079
No log 5.15 412 0.9149 0.7135 0.9149 0.9565
No log 5.175 414 1.0319 0.6740 1.0319 1.0158
No log 5.2 416 1.0948 0.6778 1.0948 1.0463
No log 5.225 418 1.0669 0.6514 1.0669 1.0329
No log 5.25 420 0.9466 0.6667 0.9466 0.9730
No log 5.275 422 0.9342 0.6709 0.9342 0.9666
No log 5.3 424 0.8961 0.6923 0.8961 0.9466
No log 5.325 426 0.8742 0.6849 0.8742 0.9350
No log 5.35 428 0.9007 0.6711 0.9007 0.9491
No log 5.375 430 0.9548 0.6497 0.9548 0.9771
No log 5.4 432 0.9381 0.6832 0.9381 0.9686
No log 5.425 434 0.9759 0.6667 0.9759 0.9879
No log 5.45 436 1.0322 0.6424 1.0322 1.0160
No log 5.475 438 1.0790 0.6429 1.0790 1.0387
No log 5.5 440 1.1373 0.6108 1.1373 1.0664
No log 5.525 442 1.2857 0.5629 1.2857 1.1339
No log 5.55 444 1.1527 0.6061 1.1527 1.0737
No log 5.575 446 0.8716 0.6624 0.8716 0.9336
No log 5.6 448 0.7479 0.7308 0.7479 0.8648
No log 5.625 450 0.6736 0.7417 0.6736 0.8207
No log 5.65 452 0.6893 0.7451 0.6893 0.8302
No log 5.675 454 0.6514 0.7532 0.6514 0.8071
No log 5.7 456 0.6465 0.7682 0.6465 0.8041
No log 5.725 458 0.6490 0.7619 0.6490 0.8056
No log 5.75 460 0.6911 0.7534 0.6911 0.8313
No log 5.775 462 0.7423 0.7376 0.7423 0.8616
No log 5.8 464 0.7820 0.7153 0.7820 0.8843
No log 5.825 466 0.8067 0.6765 0.8067 0.8982
No log 5.85 468 0.9219 0.5793 0.9219 0.9602
No log 5.875 470 1.0432 0.6013 1.0432 1.0213
No log 5.9 472 0.9515 0.6282 0.9515 0.9755
No log 5.925 474 0.8187 0.7407 0.8187 0.9048
No log 5.95 476 0.7097 0.7831 0.7097 0.8424
No log 5.975 478 0.7357 0.7758 0.7357 0.8577
No log 6.0 480 0.7609 0.7485 0.7609 0.8723
No log 6.025 482 0.8299 0.7561 0.8299 0.9110
No log 6.05 484 0.8166 0.7665 0.8166 0.9037
No log 6.075 486 0.7946 0.7665 0.7946 0.8914
No log 6.1 488 0.7933 0.7665 0.7933 0.8907
No log 6.125 490 0.7188 0.8025 0.7188 0.8478
No log 6.15 492 0.6495 0.7895 0.6495 0.8059
No log 6.175 494 0.6818 0.7671 0.6818 0.8257
No log 6.2 496 0.7152 0.7092 0.7152 0.8457
No log 6.225 498 0.6978 0.7429 0.6978 0.8353
0.4143 6.25 500 0.6833 0.7586 0.6833 0.8266
0.4143 6.275 502 0.6856 0.7703 0.6856 0.8280
0.4143 6.3 504 0.6728 0.7703 0.6728 0.8202
0.4143 6.325 506 0.6952 0.7429 0.6952 0.8338
0.4143 6.35 508 0.7066 0.7552 0.7066 0.8406
0.4143 6.375 510 0.7161 0.7785 0.7161 0.8462
0.4143 6.4 512 0.7279 0.7821 0.7279 0.8532
0.4143 6.425 514 0.7779 0.7643 0.7779 0.8820
0.4143 6.45 516 0.8811 0.7051 0.8811 0.9387
0.4143 6.475 518 0.9230 0.6795 0.9230 0.9607
0.4143 6.5 520 0.9451 0.6795 0.9451 0.9722
0.4143 6.525 522 0.8644 0.6752 0.8644 0.9298
0.4143 6.55 524 0.8347 0.725 0.8347 0.9136

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k17_task1_organization

Finetuned
(4019)
this model