ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k20_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7362
  • Qwk: 0.4968
  • Mse: 0.7362
  • Rmse: 0.8580

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0312 2 4.0813 -0.0061 4.0813 2.0202
No log 0.0625 4 3.2642 -0.0107 3.2642 1.8067
No log 0.0938 6 2.6855 -0.0367 2.6855 1.6388
No log 0.125 8 1.4433 -0.0457 1.4433 1.2014
No log 0.1562 10 1.3022 -0.0180 1.3022 1.1411
No log 0.1875 12 1.2260 0.0790 1.2260 1.1073
No log 0.2188 14 1.5237 -0.0144 1.5237 1.2344
No log 0.25 16 1.7827 0.0181 1.7827 1.3352
No log 0.2812 18 2.0241 0.0916 2.0241 1.4227
No log 0.3125 20 1.8281 0.0423 1.8281 1.3521
No log 0.3438 22 1.3743 0.0116 1.3743 1.1723
No log 0.375 24 1.2435 0.0085 1.2435 1.1151
No log 0.4062 26 1.1194 0.1086 1.1194 1.0580
No log 0.4375 28 1.0640 0.1953 1.0640 1.0315
No log 0.4688 30 1.0920 0.1858 1.0920 1.0450
No log 0.5 32 1.1134 0.1711 1.1134 1.0552
No log 0.5312 34 1.2933 0.1086 1.2933 1.1372
No log 0.5625 36 1.1165 0.1979 1.1165 1.0566
No log 0.5938 38 0.9709 0.2192 0.9709 0.9853
No log 0.625 40 0.9505 0.2448 0.9505 0.9749
No log 0.6562 42 1.0792 0.2644 1.0792 1.0389
No log 0.6875 44 1.6160 0.2059 1.6160 1.2712
No log 0.7188 46 1.9860 0.1362 1.9860 1.4093
No log 0.75 48 1.8744 0.1187 1.8744 1.3691
No log 0.7812 50 1.5828 0.1674 1.5828 1.2581
No log 0.8125 52 1.2218 0.1344 1.2218 1.1053
No log 0.8438 54 1.0027 0.2828 1.0027 1.0013
No log 0.875 56 1.0362 0.2306 1.0362 1.0179
No log 0.9062 58 1.0337 0.2163 1.0337 1.0167
No log 0.9375 60 1.0427 0.2448 1.0427 1.0211
No log 0.9688 62 1.0248 0.2755 1.0248 1.0123
No log 1.0 64 1.0250 0.3431 1.0250 1.0124
No log 1.0312 66 1.0459 0.3720 1.0459 1.0227
No log 1.0625 68 0.9425 0.2967 0.9425 0.9708
No log 1.0938 70 0.9093 0.4355 0.9093 0.9536
No log 1.125 72 0.9826 0.2692 0.9826 0.9913
No log 1.1562 74 1.1177 0.1848 1.1177 1.0572
No log 1.1875 76 1.0742 0.3328 1.0742 1.0365
No log 1.2188 78 1.1386 0.3333 1.1386 1.0671
No log 1.25 80 1.3412 0.3715 1.3412 1.1581
No log 1.2812 82 1.4205 0.3369 1.4205 1.1918
No log 1.3125 84 1.2296 0.3668 1.2296 1.1089
No log 1.3438 86 0.9793 0.4009 0.9793 0.9896
No log 1.375 88 0.8397 0.5074 0.8397 0.9164
No log 1.4062 90 0.8474 0.4843 0.8474 0.9205
No log 1.4375 92 0.8281 0.5098 0.8281 0.9100
No log 1.4688 94 0.8474 0.4368 0.8474 0.9206
No log 1.5 96 0.9561 0.4224 0.9561 0.9778
No log 1.5312 98 1.0941 0.3848 1.0941 1.0460
No log 1.5625 100 1.0303 0.4325 1.0303 1.0150
No log 1.5938 102 0.7931 0.5205 0.7931 0.8905
No log 1.625 104 0.7848 0.5067 0.7848 0.8859
No log 1.6562 106 0.8160 0.4373 0.8160 0.9033
No log 1.6875 108 0.7592 0.5017 0.7592 0.8713
No log 1.7188 110 0.7850 0.4963 0.7850 0.8860
No log 1.75 112 0.7723 0.5166 0.7723 0.8788
No log 1.7812 114 0.8020 0.4529 0.8020 0.8955
No log 1.8125 116 0.8463 0.4604 0.8463 0.9199
No log 1.8438 118 0.8727 0.4489 0.8727 0.9342
No log 1.875 120 0.7772 0.5343 0.7772 0.8816
No log 1.9062 122 0.8309 0.4657 0.8309 0.9115
No log 1.9375 124 0.8131 0.3970 0.8131 0.9017
No log 1.9688 126 0.7921 0.4893 0.7921 0.8900
No log 2.0 128 0.8198 0.5348 0.8198 0.9054
No log 2.0312 130 0.8223 0.4878 0.8223 0.9068
No log 2.0625 132 0.8266 0.4960 0.8266 0.9092
No log 2.0938 134 0.8578 0.3715 0.8578 0.9262
No log 2.125 136 0.8055 0.4329 0.8055 0.8975
No log 2.1562 138 0.8151 0.4419 0.8151 0.9028
No log 2.1875 140 0.8398 0.4538 0.8398 0.9164
No log 2.2188 142 0.8004 0.4983 0.8004 0.8947
No log 2.25 144 0.7929 0.5131 0.7929 0.8905
No log 2.2812 146 0.8476 0.5638 0.8476 0.9207
No log 2.3125 148 0.8558 0.4996 0.8558 0.9251
No log 2.3438 150 0.8795 0.4778 0.8795 0.9378
No log 2.375 152 0.8762 0.4769 0.8762 0.9361
No log 2.4062 154 0.8463 0.4771 0.8463 0.9200
No log 2.4375 156 0.8297 0.5287 0.8297 0.9109
No log 2.4688 158 0.8281 0.5176 0.8281 0.9100
No log 2.5 160 0.8220 0.5517 0.8220 0.9066
No log 2.5312 162 0.8326 0.5312 0.8326 0.9124
No log 2.5625 164 0.8608 0.5222 0.8608 0.9278
No log 2.5938 166 0.7835 0.6134 0.7835 0.8852
No log 2.625 168 0.7814 0.6068 0.7814 0.8840
No log 2.6562 170 0.7848 0.5818 0.7848 0.8859
No log 2.6875 172 0.7823 0.5818 0.7823 0.8845
No log 2.7188 174 0.7818 0.6217 0.7818 0.8842
No log 2.75 176 0.7951 0.4981 0.7951 0.8917
No log 2.7812 178 0.9116 0.4555 0.9116 0.9548
No log 2.8125 180 0.8534 0.3523 0.8534 0.9238
No log 2.8438 182 0.8482 0.4110 0.8482 0.9210
No log 2.875 184 0.8704 0.3856 0.8704 0.9330
No log 2.9062 186 0.8792 0.4161 0.8792 0.9377
No log 2.9375 188 0.8359 0.4124 0.8359 0.9143
No log 2.9688 190 0.8595 0.4335 0.8595 0.9271
No log 3.0 192 0.9152 0.4326 0.9152 0.9567
No log 3.0312 194 0.8635 0.4224 0.8635 0.9292
No log 3.0625 196 0.7899 0.5356 0.7899 0.8888
No log 3.0938 198 0.7662 0.5356 0.7662 0.8753
No log 3.125 200 0.7698 0.5342 0.7698 0.8774
No log 3.1562 202 0.7840 0.5648 0.7840 0.8854
No log 3.1875 204 0.8006 0.4867 0.8006 0.8948
No log 3.2188 206 0.7907 0.4867 0.7907 0.8892
No log 3.25 208 0.7968 0.4867 0.7968 0.8926
No log 3.2812 210 0.7965 0.4867 0.7965 0.8925
No log 3.3125 212 0.7860 0.5868 0.7860 0.8866
No log 3.3438 214 0.8105 0.5313 0.8105 0.9003
No log 3.375 216 0.9031 0.4671 0.9031 0.9503
No log 3.4062 218 0.8445 0.5737 0.8445 0.9190
No log 3.4375 220 0.8340 0.5523 0.8340 0.9133
No log 3.4688 222 0.8313 0.5462 0.8313 0.9117
No log 3.5 224 0.8679 0.4553 0.8679 0.9316
No log 3.5312 226 0.8503 0.4657 0.8503 0.9221
No log 3.5625 228 0.9179 0.4087 0.9179 0.9581
No log 3.5938 230 0.8777 0.3890 0.8777 0.9369
No log 3.625 232 0.7883 0.5146 0.7883 0.8879
No log 3.6562 234 0.8281 0.5494 0.8281 0.9100
No log 3.6875 236 0.7806 0.5797 0.7806 0.8835
No log 3.7188 238 0.7797 0.5038 0.7797 0.8830
No log 3.75 240 0.7950 0.4849 0.7950 0.8916
No log 3.7812 242 0.8149 0.4991 0.8149 0.9027
No log 3.8125 244 0.7546 0.5287 0.7546 0.8687
No log 3.8438 246 0.7430 0.5774 0.7430 0.8620
No log 3.875 248 0.7617 0.5540 0.7617 0.8728
No log 3.9062 250 0.7432 0.5121 0.7432 0.8621
No log 3.9375 252 0.7338 0.4792 0.7338 0.8566
No log 3.9688 254 0.7480 0.4659 0.7480 0.8648
No log 4.0 256 0.7659 0.4626 0.7659 0.8751
No log 4.0312 258 0.7719 0.4745 0.7719 0.8786
No log 4.0625 260 0.7595 0.4878 0.7595 0.8715
No log 4.0938 262 0.7871 0.4728 0.7871 0.8872
No log 4.125 264 0.7747 0.4869 0.7747 0.8802
No log 4.1562 266 0.7577 0.5450 0.7577 0.8704
No log 4.1875 268 0.7730 0.4594 0.7730 0.8792
No log 4.2188 270 0.8075 0.3719 0.8075 0.8986
No log 4.25 272 0.8009 0.4192 0.8009 0.8949
No log 4.2812 274 0.8211 0.5103 0.8211 0.9061
No log 4.3125 276 0.8161 0.4770 0.8161 0.9034
No log 4.3438 278 0.8178 0.3836 0.8178 0.9043
No log 4.375 280 0.8173 0.3896 0.8173 0.9040
No log 4.4062 282 0.8253 0.4527 0.8253 0.9085
No log 4.4375 284 0.8363 0.3496 0.8363 0.9145
No log 4.4688 286 0.8468 0.3428 0.8468 0.9202
No log 4.5 288 0.8348 0.5089 0.8348 0.9137
No log 4.5312 290 0.8086 0.4544 0.8086 0.8992
No log 4.5625 292 0.8155 0.4743 0.8155 0.9031
No log 4.5938 294 0.8813 0.5230 0.8813 0.9388
No log 4.625 296 0.8478 0.5042 0.8478 0.9208
No log 4.6562 298 0.8147 0.4180 0.8147 0.9026
No log 4.6875 300 0.8421 0.4174 0.8421 0.9177
No log 4.7188 302 0.8114 0.4444 0.8114 0.9008
No log 4.75 304 0.7688 0.4485 0.7688 0.8768
No log 4.7812 306 0.7544 0.4416 0.7544 0.8685
No log 4.8125 308 0.7513 0.4537 0.7513 0.8668
No log 4.8438 310 0.7446 0.3858 0.7446 0.8629
No log 4.875 312 0.7418 0.4361 0.7418 0.8613
No log 4.9062 314 0.7265 0.4503 0.7265 0.8524
No log 4.9375 316 0.7481 0.4433 0.7481 0.8649
No log 4.9688 318 0.7620 0.4312 0.7620 0.8729
No log 5.0 320 0.7492 0.4778 0.7492 0.8655
No log 5.0312 322 0.7558 0.5127 0.7558 0.8694
No log 5.0625 324 0.7601 0.4745 0.7601 0.8718
No log 5.0938 326 0.7569 0.4361 0.7569 0.8700
No log 5.125 328 0.7265 0.6043 0.7265 0.8524
No log 5.1562 330 0.7231 0.5844 0.7231 0.8503
No log 5.1875 332 0.6947 0.5131 0.6947 0.8335
No log 5.2188 334 0.6830 0.5131 0.6830 0.8264
No log 5.25 336 0.6932 0.4797 0.6932 0.8326
No log 5.2812 338 0.7050 0.4520 0.7050 0.8396
No log 5.3125 340 0.7063 0.4520 0.7063 0.8404
No log 5.3438 342 0.7191 0.4241 0.7191 0.8480
No log 5.375 344 0.7353 0.4878 0.7353 0.8575
No log 5.4062 346 0.7180 0.4912 0.7180 0.8473
No log 5.4375 348 0.7283 0.5480 0.7283 0.8534
No log 5.4688 350 0.7317 0.5450 0.7317 0.8554
No log 5.5 352 0.7172 0.4802 0.7172 0.8469
No log 5.5312 354 0.7841 0.4803 0.7841 0.8855
No log 5.5625 356 0.8646 0.4942 0.8646 0.9298
No log 5.5938 358 0.7791 0.4439 0.7791 0.8827
No log 5.625 360 0.6948 0.4660 0.6948 0.8336
No log 5.6562 362 0.7054 0.5131 0.7054 0.8399
No log 5.6875 364 0.6997 0.5503 0.6997 0.8365
No log 5.7188 366 0.7011 0.5274 0.7011 0.8373
No log 5.75 368 0.7190 0.5103 0.7190 0.8480
No log 5.7812 370 0.7098 0.4405 0.7098 0.8425
No log 5.8125 372 0.7169 0.4759 0.7169 0.8467
No log 5.8438 374 0.7342 0.5410 0.7342 0.8568
No log 5.875 376 0.7169 0.5036 0.7169 0.8467
No log 5.9062 378 0.7601 0.5542 0.7601 0.8718
No log 5.9375 380 0.7559 0.5530 0.7559 0.8694
No log 5.9688 382 0.7346 0.5979 0.7346 0.8571
No log 6.0 384 0.7417 0.6097 0.7417 0.8612
No log 6.0312 386 0.7791 0.5922 0.7791 0.8826
No log 6.0625 388 0.7184 0.5737 0.7184 0.8476
No log 6.0938 390 0.7050 0.5396 0.7050 0.8396
No log 6.125 392 0.7052 0.4834 0.7052 0.8398
No log 6.1562 394 0.7545 0.4883 0.7545 0.8686
No log 6.1875 396 0.8022 0.4797 0.8022 0.8956
No log 6.2188 398 0.8160 0.4808 0.8160 0.9033
No log 6.25 400 0.7430 0.5459 0.7430 0.8620
No log 6.2812 402 0.6882 0.5071 0.6882 0.8295
No log 6.3125 404 0.6891 0.5610 0.6891 0.8301
No log 6.3438 406 0.7275 0.5548 0.7275 0.8529
No log 6.375 408 0.7256 0.5674 0.7256 0.8518
No log 6.4062 410 0.6894 0.4927 0.6894 0.8303
No log 6.4375 412 0.6901 0.5721 0.6901 0.8307
No log 6.4688 414 0.6897 0.5593 0.6897 0.8305
No log 6.5 416 0.6824 0.4560 0.6824 0.8261
No log 6.5312 418 0.6868 0.4691 0.6868 0.8287
No log 6.5625 420 0.6864 0.5245 0.6864 0.8285
No log 6.5938 422 0.7214 0.5074 0.7214 0.8493
No log 6.625 424 0.7247 0.5074 0.7247 0.8513
No log 6.6562 426 0.6938 0.4503 0.6938 0.8330
No log 6.6875 428 0.7197 0.5463 0.7197 0.8484
No log 6.7188 430 0.7325 0.5351 0.7325 0.8559
No log 6.75 432 0.6932 0.5438 0.6932 0.8326
No log 6.7812 434 0.7059 0.5698 0.7059 0.8402
No log 6.8125 436 0.7684 0.5242 0.7684 0.8766
No log 6.8438 438 0.7521 0.5470 0.7521 0.8673
No log 6.875 440 0.7184 0.5329 0.7184 0.8476
No log 6.9062 442 0.7230 0.5224 0.7230 0.8503
No log 6.9375 444 0.7439 0.5645 0.7439 0.8625
No log 6.9688 446 0.7436 0.5431 0.7436 0.8623
No log 7.0 448 0.7480 0.5730 0.7480 0.8649
No log 7.0312 450 0.7475 0.5442 0.7475 0.8646
No log 7.0625 452 0.7405 0.5585 0.7405 0.8605
No log 7.0938 454 0.7376 0.4507 0.7376 0.8588
No log 7.125 456 0.7416 0.4507 0.7416 0.8612
No log 7.1562 458 0.7388 0.4804 0.7388 0.8595
No log 7.1875 460 0.7385 0.4804 0.7385 0.8594
No log 7.2188 462 0.7340 0.4643 0.7340 0.8567
No log 7.25 464 0.7324 0.4642 0.7324 0.8558
No log 7.2812 466 0.7284 0.4642 0.7284 0.8534
No log 7.3125 468 0.7077 0.4794 0.7077 0.8412
No log 7.3438 470 0.7320 0.5219 0.7320 0.8556
No log 7.375 472 0.9046 0.5139 0.9046 0.9511
No log 7.4062 474 0.9231 0.4526 0.9231 0.9608
No log 7.4375 476 0.7433 0.5442 0.7433 0.8622
No log 7.4688 478 0.7095 0.5359 0.7095 0.8423
No log 7.5 480 0.7366 0.5534 0.7366 0.8583
No log 7.5312 482 0.7140 0.5246 0.7140 0.8450
No log 7.5625 484 0.7255 0.5202 0.7255 0.8518
No log 7.5938 486 0.8117 0.5141 0.8117 0.9009
No log 7.625 488 0.8041 0.4784 0.8041 0.8967
No log 7.6562 490 0.7286 0.4838 0.7286 0.8536
No log 7.6875 492 0.6900 0.4807 0.6900 0.8307
No log 7.7188 494 0.6741 0.4883 0.6741 0.8210
No log 7.75 496 0.6649 0.5785 0.6649 0.8154
No log 7.7812 498 0.6984 0.5876 0.6984 0.8357
0.2888 7.8125 500 0.7981 0.5245 0.7981 0.8934
0.2888 7.8438 502 0.8137 0.5348 0.8137 0.9020
0.2888 7.875 504 0.7702 0.5128 0.7702 0.8776
0.2888 7.9062 506 0.7371 0.5674 0.7371 0.8586
0.2888 7.9375 508 0.7679 0.5125 0.7679 0.8763
0.2888 7.9688 510 0.8526 0.4903 0.8526 0.9234
0.2888 8.0 512 0.8807 0.5270 0.8807 0.9385
0.2888 8.0312 514 0.8176 0.5125 0.8176 0.9042
0.2888 8.0625 516 0.7301 0.5328 0.7301 0.8545
0.2888 8.0938 518 0.7235 0.4659 0.7235 0.8506
0.2888 8.125 520 0.7220 0.4659 0.7220 0.8497
0.2888 8.1562 522 0.7189 0.4540 0.7189 0.8479
0.2888 8.1875 524 0.7298 0.4826 0.7298 0.8543
0.2888 8.2188 526 0.7362 0.4968 0.7362 0.8580

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k20_task5_organization

Finetuned
(4019)
this model