ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k16_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6526
  • Qwk: 0.7413
  • Mse: 0.6526
  • Rmse: 0.8078

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0263 2 7.0925 -0.0056 7.0925 2.6632
No log 0.0526 4 4.6226 0.0637 4.6226 2.1500
No log 0.0789 6 4.3005 -0.0939 4.3005 2.0738
No log 0.1053 8 3.1872 0.0479 3.1872 1.7853
No log 0.1316 10 2.4290 0.0593 2.4290 1.5585
No log 0.1579 12 1.7517 0.1887 1.7517 1.3235
No log 0.1842 14 1.7255 0.2056 1.7255 1.3136
No log 0.2105 16 1.6952 0.2385 1.6952 1.3020
No log 0.2368 18 1.7727 0.2314 1.7727 1.3314
No log 0.2632 20 1.8768 0.2063 1.8768 1.3700
No log 0.2895 22 1.5760 0.3140 1.5760 1.2554
No log 0.3158 24 1.6724 0.3150 1.6724 1.2932
No log 0.3421 26 1.8448 0.3478 1.8448 1.3582
No log 0.3684 28 2.1454 0.3462 2.1454 1.4647
No log 0.3947 30 2.2317 0.35 2.2317 1.4939
No log 0.4211 32 1.9014 0.3529 1.9014 1.3789
No log 0.4474 34 1.7202 0.3673 1.7202 1.3115
No log 0.4737 36 2.0254 0.4146 2.0254 1.4232
No log 0.5 38 2.7544 0.2404 2.7544 1.6596
No log 0.5263 40 2.8142 0.2174 2.8142 1.6775
No log 0.5526 42 2.7402 0.1628 2.7402 1.6554
No log 0.5789 44 2.1522 0.2282 2.1522 1.4670
No log 0.6053 46 1.9520 0.3188 1.9520 1.3971
No log 0.6316 48 1.7304 0.3817 1.7304 1.3155
No log 0.6579 50 1.6109 0.3692 1.6109 1.2692
No log 0.6842 52 1.8453 0.3571 1.8453 1.3584
No log 0.7105 54 2.0735 0.2550 2.0735 1.4399
No log 0.7368 56 2.1260 0.25 2.1260 1.4581
No log 0.7632 58 1.7260 0.4861 1.7260 1.3138
No log 0.7895 60 1.3812 0.5455 1.3812 1.1752
No log 0.8158 62 1.0747 0.5816 1.0747 1.0367
No log 0.8421 64 1.1888 0.5915 1.1888 1.0903
No log 0.8684 66 1.3032 0.5316 1.3032 1.1416
No log 0.8947 68 1.2332 0.5412 1.2332 1.1105
No log 0.9211 70 1.2557 0.5696 1.2557 1.1206
No log 0.9474 72 1.4845 0.4789 1.4845 1.2184
No log 0.9737 74 2.6635 0.0685 2.6635 1.6320
No log 1.0 76 2.8341 -0.1493 2.8341 1.6835
No log 1.0263 78 1.7180 0.2419 1.7180 1.3107
No log 1.0526 80 0.8198 0.7143 0.8198 0.9054
No log 1.0789 82 0.9629 0.6620 0.9629 0.9813
No log 1.1053 84 1.2289 0.5921 1.2289 1.1085
No log 1.1316 86 1.1230 0.6144 1.1230 1.0597
No log 1.1579 88 0.8402 0.6575 0.8402 0.9166
No log 1.1842 90 0.8543 0.6579 0.8543 0.9243
No log 1.2105 92 1.0079 0.6410 1.0079 1.0039
No log 1.2368 94 1.1727 0.625 1.1727 1.0829
No log 1.2632 96 1.4791 0.6 1.4791 1.2162
No log 1.2895 98 1.3566 0.6257 1.3566 1.1647
No log 1.3158 100 1.0163 0.6415 1.0163 1.0081
No log 1.3421 102 0.8979 0.6479 0.8979 0.9476
No log 1.3684 104 0.9361 0.6471 0.9361 0.9675
No log 1.3947 106 0.9613 0.6269 0.9613 0.9805
No log 1.4211 108 0.9712 0.6043 0.9712 0.9855
No log 1.4474 110 1.0300 0.6197 1.0300 1.0149
No log 1.4737 112 0.9251 0.6351 0.9251 0.9618
No log 1.5 114 0.7779 0.6892 0.7779 0.8820
No log 1.5263 116 0.7660 0.7075 0.7660 0.8752
No log 1.5526 118 0.7741 0.6892 0.7741 0.8798
No log 1.5789 120 0.8463 0.7006 0.8463 0.9200
No log 1.6053 122 0.8526 0.6839 0.8526 0.9233
No log 1.6316 124 0.8024 0.6803 0.8024 0.8958
No log 1.6579 126 0.8121 0.6667 0.8121 0.9012
No log 1.6842 128 0.8581 0.6887 0.8581 0.9264
No log 1.7105 130 0.8358 0.7013 0.8358 0.9142
No log 1.7368 132 0.7909 0.7333 0.7909 0.8893
No log 1.7632 134 0.8028 0.7582 0.8028 0.8960
No log 1.7895 136 0.8308 0.7170 0.8308 0.9115
No log 1.8158 138 0.9929 0.6988 0.9929 0.9964
No log 1.8421 140 0.9398 0.6988 0.9398 0.9694
No log 1.8684 142 0.7572 0.7848 0.7572 0.8702
No log 1.8947 144 0.7515 0.7564 0.7515 0.8669
No log 1.9211 146 0.7571 0.7143 0.7571 0.8701
No log 1.9474 148 0.8221 0.7143 0.8221 0.9067
No log 1.9737 150 1.1287 0.6545 1.1287 1.0624
No log 2.0 152 1.5770 0.5587 1.5770 1.2558
No log 2.0263 154 1.5023 0.5650 1.5023 1.2257
No log 2.0526 156 0.9201 0.6883 0.9201 0.9592
No log 2.0789 158 0.8431 0.6715 0.8431 0.9182
No log 2.1053 160 0.8893 0.6412 0.8893 0.9431
No log 2.1316 162 0.9181 0.5821 0.9181 0.9582
No log 2.1579 164 1.0022 0.5224 1.0022 1.0011
No log 2.1842 166 1.0050 0.5652 1.0050 1.0025
No log 2.2105 168 0.9096 0.5455 0.9096 0.9537
No log 2.2368 170 0.8641 0.6615 0.8641 0.9296
No log 2.2632 172 0.8481 0.6615 0.8481 0.9209
No log 2.2895 174 0.7525 0.6763 0.7525 0.8675
No log 2.3158 176 0.7658 0.6933 0.7658 0.8751
No log 2.3421 178 0.8548 0.7470 0.8548 0.9246
No log 2.3684 180 0.7513 0.7453 0.7513 0.8668
No log 2.3947 182 0.7138 0.7368 0.7138 0.8449
No log 2.4211 184 0.9032 0.6377 0.9032 0.9504
No log 2.4474 186 1.0756 0.5414 1.0756 1.0371
No log 2.4737 188 0.9574 0.6111 0.9574 0.9785
No log 2.5 190 0.8921 0.7468 0.8921 0.9445
No log 2.5263 192 0.9465 0.6667 0.9465 0.9729
No log 2.5526 194 0.8927 0.6957 0.8927 0.9448
No log 2.5789 196 0.9038 0.6875 0.9038 0.9507
No log 2.6053 198 0.9004 0.6918 0.9004 0.9489
No log 2.6316 200 0.8492 0.7436 0.8492 0.9215
No log 2.6579 202 0.9857 0.6667 0.9857 0.9928
No log 2.6842 204 1.2128 0.6024 1.2128 1.1013
No log 2.7105 206 1.0309 0.6296 1.0309 1.0153
No log 2.7368 208 0.7636 0.7778 0.7636 0.8739
No log 2.7632 210 0.6609 0.7059 0.6609 0.8130
No log 2.7895 212 0.6542 0.6974 0.6542 0.8088
No log 2.8158 214 0.6204 0.7799 0.6204 0.7876
No log 2.8421 216 0.6089 0.7901 0.6089 0.7803
No log 2.8684 218 0.6221 0.7826 0.6221 0.7888
No log 2.8947 220 0.6692 0.7403 0.6692 0.8180
No log 2.9211 222 0.6908 0.75 0.6908 0.8312
No log 2.9474 224 0.7188 0.7286 0.7188 0.8478
No log 2.9737 226 0.7381 0.7286 0.7381 0.8591
No log 3.0 228 0.7321 0.6849 0.7321 0.8556
No log 3.0263 230 0.7679 0.7484 0.7679 0.8763
No log 3.0526 232 0.8047 0.7284 0.8047 0.8970
No log 3.0789 234 0.7700 0.7484 0.7700 0.8775
No log 3.1053 236 0.7470 0.72 0.7470 0.8643
No log 3.1316 238 0.7608 0.6620 0.7608 0.8722
No log 3.1579 240 0.7722 0.6522 0.7722 0.8787
No log 3.1842 242 0.7890 0.6712 0.7890 0.8882
No log 3.2105 244 0.8381 0.7051 0.8381 0.9155
No log 3.2368 246 0.8359 0.7179 0.8359 0.9143
No log 3.2632 248 0.7130 0.7123 0.7130 0.8444
No log 3.2895 250 0.6776 0.7273 0.6776 0.8232
No log 3.3158 252 0.6719 0.7194 0.6719 0.8197
No log 3.3421 254 0.6738 0.7133 0.6738 0.8209
No log 3.3684 256 0.8300 0.7051 0.8300 0.9111
No log 3.3947 258 0.9168 0.675 0.9168 0.9575
No log 3.4211 260 0.7984 0.7143 0.7984 0.8935
No log 3.4474 262 0.7047 0.7285 0.7047 0.8395
No log 3.4737 264 0.7014 0.7397 0.7014 0.8375
No log 3.5 266 0.7247 0.7285 0.7247 0.8513
No log 3.5263 268 0.7585 0.7342 0.7585 0.8709
No log 3.5526 270 0.8152 0.7044 0.8152 0.9029
No log 3.5789 272 0.8934 0.6918 0.8934 0.9452
No log 3.6053 274 0.8906 0.6918 0.8906 0.9437
No log 3.6316 276 0.8361 0.7284 0.8361 0.9144
No log 3.6579 278 0.8065 0.7394 0.8065 0.8981
No log 3.6842 280 0.8477 0.7317 0.8477 0.9207
No log 3.7105 282 0.7637 0.7179 0.7637 0.8739
No log 3.7368 284 0.7100 0.7105 0.7100 0.8426
No log 3.7632 286 0.7665 0.6842 0.7665 0.8755
No log 3.7895 288 0.8701 0.6711 0.8701 0.9328
No log 3.8158 290 0.8509 0.6753 0.8509 0.9225
No log 3.8421 292 0.7309 0.7368 0.7309 0.8549
No log 3.8684 294 0.7813 0.7285 0.7813 0.8839
No log 3.8947 296 0.7854 0.7547 0.7854 0.8862
No log 3.9211 298 0.7712 0.7349 0.7712 0.8782
No log 3.9474 300 0.7925 0.7514 0.7925 0.8902
No log 3.9737 302 0.7619 0.7647 0.7619 0.8729
No log 4.0 304 0.7351 0.7545 0.7351 0.8574
No log 4.0263 306 0.7625 0.7362 0.7625 0.8732
No log 4.0526 308 0.8134 0.7143 0.8134 0.9019
No log 4.0789 310 0.9570 0.6452 0.9570 0.9783
No log 4.1053 312 0.9519 0.64 0.9519 0.9757
No log 4.1316 314 0.8533 0.7020 0.8533 0.9238
No log 4.1579 316 0.8352 0.7020 0.8352 0.9139
No log 4.1842 318 0.8669 0.6711 0.8669 0.9311
No log 4.2105 320 0.8566 0.6438 0.8566 0.9255
No log 4.2368 322 0.8496 0.6714 0.8496 0.9217
No log 4.2632 324 0.8492 0.6906 0.8492 0.9215
No log 4.2895 326 0.8216 0.6906 0.8216 0.9064
No log 4.3158 328 0.8113 0.6803 0.8113 0.9007
No log 4.3421 330 0.8040 0.6712 0.8040 0.8967
No log 4.3684 332 0.7443 0.72 0.7443 0.8627
No log 4.3947 334 0.6440 0.7582 0.6440 0.8025
No log 4.4211 336 0.6152 0.7742 0.6152 0.7844
No log 4.4474 338 0.6033 0.7742 0.6033 0.7767
No log 4.4737 340 0.6368 0.7582 0.6368 0.7980
No log 4.5 342 0.7925 0.7531 0.7925 0.8902
No log 4.5263 344 0.9516 0.6626 0.9516 0.9755
No log 4.5526 346 0.8732 0.7081 0.8732 0.9345
No log 4.5789 348 0.7790 0.75 0.7790 0.8826
No log 4.6053 350 0.7967 0.7308 0.7967 0.8926
No log 4.6316 352 0.9276 0.6538 0.9276 0.9631
No log 4.6579 354 1.0139 0.6625 1.0139 1.0069
No log 4.6842 356 0.8622 0.6581 0.8622 0.9286
No log 4.7105 358 0.7186 0.7436 0.7186 0.8477
No log 4.7368 360 0.6618 0.7619 0.6618 0.8135
No log 4.7632 362 0.6719 0.7482 0.6719 0.8197
No log 4.7895 364 0.6910 0.7234 0.6910 0.8313
No log 4.8158 366 0.7744 0.7248 0.7744 0.8800
No log 4.8421 368 0.8455 0.7273 0.8455 0.9195
No log 4.8684 370 0.7206 0.75 0.7206 0.8489
No log 4.8947 372 0.6129 0.7568 0.6129 0.7829
No log 4.9211 374 0.6201 0.7260 0.6201 0.7875
No log 4.9474 376 0.6143 0.7260 0.6143 0.7838
No log 4.9737 378 0.5875 0.7632 0.5875 0.7665
No log 5.0 380 0.6569 0.7771 0.6569 0.8105
No log 5.0263 382 0.6898 0.7799 0.6898 0.8305
No log 5.0526 384 0.6655 0.7871 0.6655 0.8158
No log 5.0789 386 0.6250 0.7662 0.6250 0.7906
No log 5.1053 388 0.5908 0.7662 0.5908 0.7686
No log 5.1316 390 0.5813 0.7733 0.5813 0.7624
No log 5.1579 392 0.5717 0.7733 0.5717 0.7561
No log 5.1842 394 0.5993 0.7413 0.5993 0.7742
No log 5.2105 396 0.6373 0.7413 0.6373 0.7983
No log 5.2368 398 0.6208 0.7413 0.6208 0.7879
No log 5.2632 400 0.6289 0.7682 0.6289 0.7930
No log 5.2895 402 0.6976 0.7531 0.6976 0.8352
No log 5.3158 404 0.6950 0.7654 0.6950 0.8336
No log 5.3421 406 0.6725 0.7564 0.6725 0.8201
No log 5.3684 408 0.6878 0.7403 0.6878 0.8293
No log 5.3947 410 0.6848 0.7355 0.6848 0.8275
No log 5.4211 412 0.6481 0.7403 0.6481 0.8051
No log 5.4474 414 0.6516 0.7484 0.6516 0.8072
No log 5.4737 416 0.6712 0.7643 0.6712 0.8193
No log 5.5 418 0.7190 0.7389 0.7190 0.8479
No log 5.5263 420 0.8076 0.7439 0.8076 0.8987
No log 5.5526 422 0.8649 0.7399 0.8649 0.9300
No log 5.5789 424 0.8062 0.7619 0.8062 0.8979
No log 5.6053 426 0.6881 0.7578 0.6881 0.8295
No log 5.6316 428 0.6111 0.7821 0.6111 0.7818
No log 5.6579 430 0.6338 0.7568 0.6338 0.7961
No log 5.6842 432 0.7078 0.7285 0.7078 0.8413
No log 5.7105 434 0.7302 0.7417 0.7302 0.8545
No log 5.7368 436 0.6474 0.7273 0.6474 0.8046
No log 5.7632 438 0.6118 0.7338 0.6118 0.7822
No log 5.7895 440 0.6438 0.7101 0.6438 0.8024
No log 5.8158 442 0.7040 0.7007 0.7040 0.8390
No log 5.8421 444 0.6842 0.7206 0.6842 0.8272
No log 5.8684 446 0.6726 0.7639 0.6726 0.8201
No log 5.8947 448 0.7031 0.7619 0.7031 0.8385
No log 5.9211 450 0.7175 0.7432 0.7175 0.8471
No log 5.9474 452 0.7016 0.7397 0.7016 0.8376
No log 5.9737 454 0.6857 0.7483 0.6857 0.8281
No log 6.0 456 0.6930 0.7632 0.6930 0.8324
No log 6.0263 458 0.7034 0.7582 0.7034 0.8387
No log 6.0526 460 0.6484 0.7712 0.6484 0.8052
No log 6.0789 462 0.5982 0.7703 0.5982 0.7734
No log 6.1053 464 0.5892 0.7733 0.5892 0.7676
No log 6.1316 466 0.6044 0.7898 0.6044 0.7774
No log 6.1579 468 0.6743 0.7389 0.6743 0.8212
No log 6.1842 470 0.7285 0.7532 0.7285 0.8535
No log 6.2105 472 0.7557 0.7152 0.7557 0.8693
No log 6.2368 474 0.7335 0.7260 0.7335 0.8565
No log 6.2632 476 0.7114 0.7619 0.7114 0.8434
No log 6.2895 478 0.6883 0.7733 0.6883 0.8297
No log 6.3158 480 0.6297 0.7875 0.6297 0.7936
No log 6.3421 482 0.6568 0.7805 0.6568 0.8104
No log 6.3684 484 0.7766 0.7456 0.7766 0.8812
No log 6.3947 486 0.7285 0.7381 0.7285 0.8535
No log 6.4211 488 0.5908 0.7799 0.5908 0.7686
No log 6.4474 490 0.5780 0.7703 0.5780 0.7603
No log 6.4737 492 0.6107 0.7703 0.6107 0.7814
No log 6.5 494 0.6565 0.7451 0.6565 0.8102
No log 6.5263 496 0.7750 0.7261 0.7750 0.8803
No log 6.5526 498 0.8224 0.7205 0.8224 0.9069
0.3767 6.5789 500 0.7128 0.7547 0.7128 0.8443
0.3767 6.6053 502 0.6177 0.7947 0.6177 0.7860
0.3767 6.6316 504 0.6287 0.7838 0.6287 0.7929
0.3767 6.6579 506 0.6432 0.7895 0.6432 0.8020
0.3767 6.6842 508 0.7276 0.7308 0.7276 0.8530
0.3767 6.7105 510 0.7911 0.7308 0.7911 0.8894
0.3767 6.7368 512 0.7628 0.7027 0.7628 0.8734
0.3767 6.7632 514 0.6920 0.7534 0.6920 0.8319
0.3767 6.7895 516 0.6526 0.7413 0.6526 0.8078

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k16_task1_organization

Finetuned
(4019)
this model