ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k15_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7493
  • Qwk: -0.0532
  • Mse: 0.7493
  • Rmse: 0.8656

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0260 2 3.8138 0.0104 3.8138 1.9529
No log 0.0519 4 2.2154 -0.0498 2.2154 1.4884
No log 0.0779 6 1.3391 0.0 1.3391 1.1572
No log 0.1039 8 1.0894 -0.0345 1.0894 1.0438
No log 0.1299 10 1.4289 0.0098 1.4289 1.1954
No log 0.1558 12 0.9306 -0.0504 0.9306 0.9647
No log 0.1818 14 0.6954 -0.0035 0.6954 0.8339
No log 0.2078 16 0.6929 -0.0035 0.6929 0.8324
No log 0.2338 18 0.7242 0.0416 0.7242 0.8510
No log 0.2597 20 0.6861 0.0506 0.6861 0.8283
No log 0.2857 22 0.6886 0.0 0.6886 0.8298
No log 0.3117 24 0.6792 0.0 0.6792 0.8242
No log 0.3377 26 0.7626 0.1047 0.7626 0.8733
No log 0.3636 28 0.9213 -0.0837 0.9213 0.9598
No log 0.3896 30 0.8146 -0.0506 0.8146 0.9026
No log 0.4156 32 0.8843 -0.1303 0.8843 0.9404
No log 0.4416 34 0.9314 0.0768 0.9314 0.9651
No log 0.4675 36 0.9769 0.0912 0.9769 0.9884
No log 0.4935 38 0.9996 0.1546 0.9996 0.9998
No log 0.5195 40 1.0869 -0.0842 1.0869 1.0425
No log 0.5455 42 1.0399 0.0045 1.0399 1.0198
No log 0.5714 44 0.9904 0.0559 0.9904 0.9952
No log 0.5974 46 1.2796 0.1106 1.2796 1.1312
No log 0.6234 48 1.3511 0.0596 1.3511 1.1624
No log 0.6494 50 0.9263 0.0069 0.9263 0.9624
No log 0.6753 52 1.3873 0.0843 1.3873 1.1778
No log 0.7013 54 1.3119 0.0390 1.3119 1.1454
No log 0.7273 56 0.7638 0.0205 0.7638 0.8740
No log 0.7532 58 1.0075 0.0196 1.0075 1.0037
No log 0.7792 60 0.9809 0.0147 0.9809 0.9904
No log 0.8052 62 0.7732 0.0393 0.7732 0.8793
No log 0.8312 64 1.0124 0.0596 1.0124 1.0062
No log 0.8571 66 0.9489 0.0496 0.9489 0.9741
No log 0.8831 68 1.0472 0.0707 1.0472 1.0233
No log 0.9091 70 1.0365 0.1249 1.0365 1.0181
No log 0.9351 72 1.0065 0.0871 1.0065 1.0033
No log 0.9610 74 1.0068 -0.0320 1.0068 1.0034
No log 0.9870 76 0.9106 -0.0178 0.9106 0.9543
No log 1.0130 78 0.9065 0.1002 0.9065 0.9521
No log 1.0390 80 0.8352 0.2080 0.8352 0.9139
No log 1.0649 82 0.8462 0.0959 0.8462 0.9199
No log 1.0909 84 0.8211 0.1633 0.8211 0.9061
No log 1.1169 86 0.8721 0.1734 0.8721 0.9339
No log 1.1429 88 0.8848 0.1313 0.8848 0.9406
No log 1.1688 90 0.9992 0.0619 0.9992 0.9996
No log 1.1948 92 1.0727 0.0619 1.0727 1.0357
No log 1.2208 94 0.9375 0.1854 0.9375 0.9683
No log 1.2468 96 0.9827 0.0730 0.9827 0.9913
No log 1.2727 98 0.8248 0.1393 0.8248 0.9082
No log 1.2987 100 0.7412 0.1304 0.7412 0.8609
No log 1.3247 102 0.7197 0.1304 0.7197 0.8483
No log 1.3506 104 0.7110 0.1807 0.7110 0.8432
No log 1.3766 106 0.7674 0.0639 0.7674 0.8760
No log 1.4026 108 0.7476 0.1311 0.7476 0.8646
No log 1.4286 110 0.8687 0.1078 0.8687 0.9320
No log 1.4545 112 0.9666 0.1042 0.9666 0.9832
No log 1.4805 114 0.9576 0.1626 0.9576 0.9786
No log 1.5065 116 0.9921 0.1626 0.9921 0.9961
No log 1.5325 118 1.0625 0.0063 1.0625 1.0308
No log 1.5584 120 1.3877 0.0285 1.3877 1.1780
No log 1.5844 122 1.2962 -0.0327 1.2962 1.1385
No log 1.6104 124 0.9704 -0.0062 0.9704 0.9851
No log 1.6364 126 1.0054 0.0288 1.0054 1.0027
No log 1.6623 128 0.8994 0.0725 0.8994 0.9483
No log 1.6883 130 0.9452 0.1419 0.9452 0.9722
No log 1.7143 132 0.8450 -0.0054 0.8450 0.9192
No log 1.7403 134 0.8886 -0.0355 0.8886 0.9427
No log 1.7662 136 0.9387 -0.1206 0.9387 0.9689
No log 1.7922 138 0.8899 0.0460 0.8899 0.9433
No log 1.8182 140 0.9194 0.0220 0.9194 0.9589
No log 1.8442 142 0.9478 -0.1501 0.9478 0.9736
No log 1.8701 144 1.0207 -0.1212 1.0207 1.0103
No log 1.8961 146 0.8871 0.0898 0.8871 0.9419
No log 1.9221 148 1.1082 -0.0360 1.1082 1.0527
No log 1.9481 150 1.0565 -0.0142 1.0565 1.0279
No log 1.9740 152 0.8089 -0.0427 0.8089 0.8994
No log 2.0 154 0.8782 -0.0842 0.8782 0.9371
No log 2.0260 156 0.8499 0.0071 0.8499 0.9219
No log 2.0519 158 0.7106 0.0416 0.7106 0.8429
No log 2.0779 160 0.7917 -0.0738 0.7917 0.8898
No log 2.1039 162 0.9907 0.0280 0.9907 0.9953
No log 2.1299 164 0.9248 0.0789 0.9248 0.9616
No log 2.1558 166 0.8119 0.1585 0.8119 0.9011
No log 2.1818 168 0.9255 0.1065 0.9255 0.9620
No log 2.2078 170 0.9786 0.0224 0.9786 0.9893
No log 2.2338 172 0.7957 0.1495 0.7957 0.8920
No log 2.2597 174 0.8590 -0.0059 0.8590 0.9268
No log 2.2857 176 1.0435 0.0516 1.0435 1.0215
No log 2.3117 178 0.8865 0.1039 0.8865 0.9415
No log 2.3377 180 0.8540 0.0876 0.8540 0.9241
No log 2.3636 182 0.9272 0.0659 0.9272 0.9629
No log 2.3896 184 0.7852 0.1292 0.7852 0.8861
No log 2.4156 186 0.9758 0.0415 0.9758 0.9878
No log 2.4416 188 1.1032 0.0525 1.1032 1.0503
No log 2.4675 190 0.8631 0.0665 0.8631 0.9290
No log 2.4935 192 0.8396 0.1633 0.8396 0.9163
No log 2.5195 194 1.0152 0.0164 1.0152 1.0076
No log 2.5455 196 0.8554 0.1582 0.8554 0.9249
No log 2.5714 198 0.7767 0.0985 0.7767 0.8813
No log 2.5974 200 0.8507 0.0647 0.8507 0.9223
No log 2.6234 202 0.8155 0.0200 0.8155 0.9031
No log 2.6494 204 0.7697 0.1292 0.7697 0.8773
No log 2.6753 206 0.9331 0.1430 0.9331 0.9660
No log 2.7013 208 0.9232 0.1430 0.9232 0.9608
No log 2.7273 210 0.8437 0.1231 0.8437 0.9185
No log 2.7532 212 0.7452 0.0828 0.7452 0.8632
No log 2.7792 214 0.7743 0.0058 0.7743 0.8799
No log 2.8052 216 0.7515 0.0058 0.7515 0.8669
No log 2.8312 218 0.7620 0.0680 0.7620 0.8729
No log 2.8571 220 0.8329 0.1449 0.8329 0.9127
No log 2.8831 222 0.9498 0.0636 0.9498 0.9746
No log 2.9091 224 0.8049 0.1431 0.8049 0.8972
No log 2.9351 226 0.7504 0.1404 0.7504 0.8663
No log 2.9610 228 0.8282 0.0595 0.8282 0.9101
No log 2.9870 230 0.8194 0.0570 0.8194 0.9052
No log 3.0130 232 0.8571 0.1498 0.8571 0.9258
No log 3.0390 234 0.8990 0.0856 0.8990 0.9482
No log 3.0649 236 0.8244 0.1287 0.8244 0.9080
No log 3.0909 238 0.8623 0.0157 0.8623 0.9286
No log 3.1169 240 0.8054 0.0089 0.8054 0.8974
No log 3.1429 242 0.7267 -0.0560 0.7267 0.8524
No log 3.1688 244 0.7707 0.0723 0.7707 0.8779
No log 3.1948 246 0.7638 0.0269 0.7638 0.8740
No log 3.2208 248 0.7744 0.0303 0.7744 0.8800
No log 3.2468 250 0.8035 0.0804 0.8035 0.8964
No log 3.2727 252 0.8314 0.0856 0.8314 0.9118
No log 3.2987 254 0.8467 0.1267 0.8467 0.9202
No log 3.3247 256 0.8639 0.1050 0.8639 0.9295
No log 3.3506 258 0.8652 0.1224 0.8652 0.9302
No log 3.3766 260 0.8560 0.0488 0.8560 0.9252
No log 3.4026 262 0.8201 0.1144 0.8201 0.9056
No log 3.4286 264 0.8230 0.0956 0.8230 0.9072
No log 3.4545 266 0.7953 0.0956 0.7953 0.8918
No log 3.4805 268 0.7661 -0.0118 0.7661 0.8753
No log 3.5065 270 0.7994 -0.0976 0.7994 0.8941
No log 3.5325 272 0.8652 0.0200 0.8652 0.9302
No log 3.5584 274 0.8321 -0.0056 0.8321 0.9122
No log 3.5844 276 0.9237 0.0799 0.9237 0.9611
No log 3.6104 278 0.8491 0.0956 0.8491 0.9215
No log 3.6364 280 0.7674 -0.0976 0.7674 0.8760
No log 3.6623 282 0.7750 -0.0449 0.7750 0.8804
No log 3.6883 284 0.7394 -0.1094 0.7394 0.8599
No log 3.7143 286 0.7382 -0.0627 0.7382 0.8592
No log 3.7403 288 0.7621 -0.0560 0.7621 0.8730
No log 3.7662 290 0.7815 0.0 0.7815 0.8840
No log 3.7922 292 0.7967 0.0257 0.7967 0.8926
No log 3.8182 294 0.7940 -0.0170 0.7940 0.8910
No log 3.8442 296 0.7873 -0.0118 0.7873 0.8873
No log 3.8701 298 0.8034 -0.0506 0.8034 0.8963
No log 3.8961 300 0.8151 0.0056 0.8151 0.9029
No log 3.9221 302 0.8592 0.0551 0.8592 0.9269
No log 3.9481 304 0.9025 0.0497 0.9025 0.9500
No log 3.9740 306 0.9202 0.1612 0.9202 0.9593
No log 4.0 308 0.9457 0.1661 0.9457 0.9725
No log 4.0260 310 0.8771 0.0608 0.8771 0.9365
No log 4.0519 312 0.9516 0.0989 0.9516 0.9755
No log 4.0779 314 0.9624 0.0576 0.9624 0.9810
No log 4.1039 316 0.7921 0.1440 0.7921 0.8900
No log 4.1299 318 0.7261 0.1758 0.7261 0.8521
No log 4.1558 320 0.7228 0.1758 0.7228 0.8502
No log 4.1818 322 0.7723 0.1627 0.7723 0.8788
No log 4.2078 324 0.8260 0.1003 0.8260 0.9088
No log 4.2338 326 0.8729 0.1006 0.8729 0.9343
No log 4.2597 328 0.9758 -0.0336 0.9758 0.9878
No log 4.2857 330 0.9573 0.0156 0.9573 0.9784
No log 4.3117 332 0.9558 0.1439 0.9558 0.9776
No log 4.3377 334 0.9761 0.0448 0.9761 0.9880
No log 4.3636 336 0.8796 0.1203 0.8796 0.9379
No log 4.3896 338 0.8253 0.1541 0.8253 0.9085
No log 4.4156 340 0.7971 0.2009 0.7971 0.8928
No log 4.4416 342 0.7588 0.0460 0.7588 0.8711
No log 4.4675 344 0.7503 0.0964 0.7503 0.8662
No log 4.4935 346 0.7572 0.1758 0.7572 0.8702
No log 4.5195 348 0.7490 0.1902 0.7490 0.8655
No log 4.5455 350 0.7759 -0.0513 0.7759 0.8808
No log 4.5714 352 0.8883 0.0220 0.8883 0.9425
No log 4.5974 354 0.8578 0.0179 0.8578 0.9262
No log 4.6234 356 0.8179 0.1475 0.8179 0.9044
No log 4.6494 358 0.8601 0.0920 0.8601 0.9274
No log 4.6753 360 0.8483 0.1049 0.8483 0.9210
No log 4.7013 362 0.8555 0.1423 0.8555 0.9250
No log 4.7273 364 0.8208 0.1144 0.8208 0.9060
No log 4.7532 366 0.7884 0.1371 0.7884 0.8879
No log 4.7792 368 0.8105 0.1254 0.8105 0.9003
No log 4.8052 370 0.8310 0.1240 0.8310 0.9116
No log 4.8312 372 0.8957 -0.0014 0.8957 0.9464
No log 4.8571 374 0.9225 -0.0014 0.9225 0.9605
No log 4.8831 376 0.9000 0.0307 0.9000 0.9487
No log 4.9091 378 0.8375 0.1192 0.8375 0.9151
No log 4.9351 380 0.8305 0.1148 0.8305 0.9113
No log 4.9610 382 0.8295 0.1148 0.8295 0.9108
No log 4.9870 384 0.8252 0.0732 0.8252 0.9084
No log 5.0130 386 0.8419 -0.0314 0.8419 0.9175
No log 5.0390 388 0.8428 -0.0314 0.8428 0.9181
No log 5.0649 390 0.8168 -0.0506 0.8168 0.9038
No log 5.0909 392 0.8105 -0.0086 0.8105 0.9003
No log 5.1169 394 0.8033 0.0085 0.8033 0.8963
No log 5.1429 396 0.7614 -0.0096 0.7614 0.8726
No log 5.1688 398 0.7358 0.0460 0.7358 0.8578
No log 5.1948 400 0.7753 0.0714 0.7753 0.8805
No log 5.2208 402 0.7874 0.0714 0.7874 0.8873
No log 5.2468 404 0.7467 0.1379 0.7467 0.8641
No log 5.2727 406 0.7480 0.0513 0.7480 0.8648
No log 5.2987 408 0.7620 0.0807 0.7620 0.8729
No log 5.3247 410 0.7905 0.1259 0.7905 0.8891
No log 5.3506 412 0.8135 0.0410 0.8135 0.9019
No log 5.3766 414 0.8924 -0.0522 0.8924 0.9447
No log 5.4026 416 0.8705 -0.0375 0.8705 0.9330
No log 5.4286 418 0.9958 0.0309 0.9958 0.9979
No log 5.4545 420 1.0983 0.0233 1.0983 1.0480
No log 5.4805 422 0.9599 0.0336 0.9599 0.9797
No log 5.5065 424 0.7939 0.0768 0.7939 0.8910
No log 5.5325 426 0.7811 -0.0469 0.7811 0.8838
No log 5.5584 428 0.7658 -0.0560 0.7658 0.8751
No log 5.5844 430 0.7849 0.1691 0.7849 0.8860
No log 5.6104 432 0.9050 -0.0008 0.9050 0.9513
No log 5.6364 434 0.9092 0.0363 0.9092 0.9535
No log 5.6623 436 0.8373 0.1318 0.8373 0.9151
No log 5.6883 438 0.8502 0.0749 0.8502 0.9220
No log 5.7143 440 0.8552 0.0289 0.8552 0.9248
No log 5.7403 442 0.8325 -0.0138 0.8325 0.9124
No log 5.7662 444 0.8121 0.1199 0.8121 0.9012
No log 5.7922 446 0.7941 0.1691 0.7941 0.8911
No log 5.8182 448 0.8065 0.1691 0.8065 0.8981
No log 5.8442 450 0.8224 0.1047 0.8224 0.9068
No log 5.8701 452 0.8059 0.1691 0.8059 0.8977
No log 5.8961 454 0.7752 0.1259 0.7752 0.8805
No log 5.9221 456 0.7946 0.1691 0.7946 0.8914
No log 5.9481 458 0.8061 0.1691 0.8061 0.8978
No log 5.9740 460 0.8766 -0.0365 0.8766 0.9363
No log 6.0 462 0.9782 0.0134 0.9782 0.9890
No log 6.0260 464 0.9065 -0.0035 0.9065 0.9521
No log 6.0519 466 0.8183 0.1691 0.8183 0.9046
No log 6.0779 468 0.9156 0.0362 0.9156 0.9569
No log 6.1039 470 0.8854 0.1342 0.8854 0.9409
No log 6.1299 472 0.7465 0.1259 0.7465 0.8640
No log 6.1558 474 0.7613 0.0 0.7613 0.8725
No log 6.1818 476 0.7656 -0.0446 0.7656 0.8750
No log 6.2078 478 0.7480 0.1259 0.7480 0.8648
No log 6.2338 480 0.8022 0.1627 0.8022 0.8956
No log 6.2597 482 0.7981 0.1627 0.7981 0.8934
No log 6.2857 484 0.8012 -0.0462 0.8012 0.8951
No log 6.3117 486 0.9166 0.0673 0.9166 0.9574
No log 6.3377 488 0.8448 -0.0159 0.8448 0.9191
No log 6.3636 490 0.7299 0.0821 0.7299 0.8544
No log 6.3896 492 0.7549 0.1202 0.7549 0.8688
No log 6.4156 494 0.7883 0.1627 0.7883 0.8878
No log 6.4416 496 0.7563 0.1259 0.7563 0.8697
No log 6.4675 498 0.7749 0.0394 0.7749 0.8803
0.2892 6.4935 500 0.7735 0.0357 0.7735 0.8795
0.2892 6.5195 502 0.7908 -0.0059 0.7908 0.8893
0.2892 6.5455 504 0.7880 0.0394 0.7880 0.8877
0.2892 6.5714 506 0.7810 0.0394 0.7810 0.8837
0.2892 6.5974 508 0.7665 0.1740 0.7665 0.8755
0.2892 6.6234 510 0.8365 0.1048 0.8365 0.9146
0.2892 6.6494 512 0.8018 0.1627 0.8018 0.8954
0.2892 6.6753 514 0.7345 0.1199 0.7345 0.8570
0.2892 6.7013 516 0.7307 0.1371 0.7307 0.8548
0.2892 6.7273 518 0.7745 -0.0506 0.7745 0.8800
0.2892 6.7532 520 0.8002 -0.0082 0.8002 0.8946
0.2892 6.7792 522 0.8248 0.1095 0.8248 0.9082
0.2892 6.8052 524 0.9060 0.0435 0.9060 0.9518
0.2892 6.8312 526 0.8598 0.1423 0.8598 0.9273
0.2892 6.8571 528 0.8414 -0.0557 0.8414 0.9173
0.2892 6.8831 530 0.8141 -0.0082 0.8141 0.9023
0.2892 6.9091 532 0.8403 -0.0277 0.8403 0.9167
0.2892 6.9351 534 0.8468 -0.0187 0.8468 0.9202
0.2892 6.9610 536 0.7493 -0.0532 0.7493 0.8656

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k15_task3_organization

Finetuned
(4019)
this model