ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8154
  • Qwk: 0.7226
  • Mse: 0.8154
  • Rmse: 0.9030

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 7.0187 -0.0055 7.0187 2.6493
No log 0.1379 4 4.2530 0.0684 4.2530 2.0623
No log 0.2069 6 3.3286 -0.0432 3.3286 1.8244
No log 0.2759 8 4.9756 -0.0591 4.9756 2.2306
No log 0.3448 10 4.5622 -0.0702 4.5622 2.1359
No log 0.4138 12 2.5120 0.1224 2.5120 1.5849
No log 0.4828 14 2.1957 0.0882 2.1957 1.4818
No log 0.5517 16 2.3582 0.0142 2.3582 1.5357
No log 0.6207 18 2.7105 0.0261 2.7105 1.6464
No log 0.6897 20 2.6846 0.0 2.6846 1.6385
No log 0.7586 22 2.6690 -0.0134 2.6690 1.6337
No log 0.8276 24 2.4597 0.0 2.4597 1.5684
No log 0.8966 26 1.9648 0.3115 1.9648 1.4017
No log 0.9655 28 1.6930 0.3103 1.6930 1.3011
No log 1.0345 30 1.6116 0.3103 1.6116 1.2695
No log 1.1034 32 1.7461 0.3438 1.7461 1.3214
No log 1.1724 34 2.3913 0.2436 2.3913 1.5464
No log 1.2414 36 2.8247 0.1829 2.8247 1.6807
No log 1.3103 38 2.2383 0.2329 2.2383 1.4961
No log 1.3793 40 1.8919 0.3308 1.8919 1.3754
No log 1.4483 42 1.7878 0.3308 1.7878 1.3371
No log 1.5172 44 1.9417 0.2754 1.9417 1.3935
No log 1.5862 46 1.8201 0.3066 1.8201 1.3491
No log 1.6552 48 1.8874 0.2754 1.8874 1.3738
No log 1.7241 50 1.9415 0.2917 1.9415 1.3934
No log 1.7931 52 2.0607 0.3137 2.0607 1.4355
No log 1.8621 54 2.1054 0.3293 2.1054 1.4510
No log 1.9310 56 2.5931 0.2951 2.5931 1.6103
No log 2.0 58 2.4507 0.3077 2.4507 1.5655
No log 2.0690 60 1.8230 0.4 1.8230 1.3502
No log 2.1379 62 1.4287 0.4031 1.4287 1.1953
No log 2.2069 64 1.4282 0.4094 1.4282 1.1951
No log 2.2759 66 1.4069 0.4286 1.4069 1.1861
No log 2.3448 68 1.4073 0.4409 1.4073 1.1863
No log 2.4138 70 1.4291 0.4930 1.4291 1.1954
No log 2.4828 72 1.2389 0.4722 1.2389 1.1131
No log 2.5517 74 1.0493 0.5507 1.0493 1.0244
No log 2.6207 76 0.9983 0.5672 0.9983 0.9991
No log 2.6897 78 1.0374 0.5588 1.0374 1.0185
No log 2.7586 80 1.1696 0.5070 1.1696 1.0815
No log 2.8276 82 1.2804 0.4714 1.2804 1.1315
No log 2.8966 84 1.4641 0.4348 1.4641 1.2100
No log 2.9655 86 1.4472 0.4643 1.4472 1.2030
No log 3.0345 88 1.4946 0.4606 1.4946 1.2225
No log 3.1034 90 1.5480 0.4944 1.5480 1.2442
No log 3.1724 92 1.3597 0.4524 1.3597 1.1661
No log 3.2414 94 1.1660 0.6135 1.1660 1.0798
No log 3.3103 96 1.0117 0.6624 1.0117 1.0058
No log 3.3793 98 1.0269 0.6971 1.0269 1.0134
No log 3.4483 100 0.9695 0.6860 0.9695 0.9846
No log 3.5172 102 0.8902 0.6667 0.8902 0.9435
No log 3.5862 104 0.8374 0.7333 0.8374 0.9151
No log 3.6552 106 0.8005 0.7226 0.8005 0.8947
No log 3.7241 108 0.8473 0.7375 0.8473 0.9205
No log 3.7931 110 0.8760 0.7407 0.8760 0.9360
No log 3.8621 112 0.9109 0.6667 0.9109 0.9544
No log 3.9310 114 0.9882 0.6225 0.9882 0.9941
No log 4.0 116 0.9843 0.6623 0.9843 0.9921
No log 4.0690 118 0.9823 0.6788 0.9823 0.9911
No log 4.1379 120 0.9051 0.6879 0.9051 0.9513
No log 4.2069 122 0.9240 0.6197 0.9240 0.9612
No log 4.2759 124 0.9535 0.6897 0.9535 0.9765
No log 4.3448 126 0.8838 0.7042 0.8838 0.9401
No log 4.4138 128 1.0769 0.6584 1.0769 1.0377
No log 4.4828 130 1.6810 0.4813 1.6810 1.2965
No log 4.5517 132 1.7386 0.4947 1.7386 1.3186
No log 4.6207 134 1.1787 0.5882 1.1787 1.0857
No log 4.6897 136 0.9681 0.6093 0.9681 0.9839
No log 4.7586 138 1.0042 0.625 1.0042 1.0021
No log 4.8276 140 0.9728 0.6197 0.9728 0.9863
No log 4.8966 142 0.7867 0.7595 0.7867 0.8869
No log 4.9655 144 0.7340 0.7389 0.7340 0.8567
No log 5.0345 146 0.7179 0.72 0.7179 0.8473
No log 5.1034 148 0.7041 0.7285 0.7041 0.8391
No log 5.1724 150 0.7697 0.7451 0.7697 0.8773
No log 5.2414 152 0.9919 0.6761 0.9919 0.9960
No log 5.3103 154 1.1585 0.5109 1.1585 1.0764
No log 5.3793 156 1.1372 0.5493 1.1372 1.0664
No log 5.4483 158 1.1191 0.5882 1.1191 1.0579
No log 5.5172 160 0.9994 0.6410 0.9994 0.9997
No log 5.5862 162 0.8691 0.7051 0.8691 0.9323
No log 5.6552 164 0.7613 0.7368 0.7613 0.8725
No log 5.7241 166 0.7581 0.75 0.7581 0.8707
No log 5.7931 168 0.7969 0.7297 0.7969 0.8927
No log 5.8621 170 0.8952 0.6338 0.8952 0.9461
No log 5.9310 172 0.9201 0.6099 0.9201 0.9592
No log 6.0 174 0.8751 0.6667 0.8751 0.9355
No log 6.0690 176 0.8878 0.7170 0.8878 0.9423
No log 6.1379 178 0.8160 0.7329 0.8160 0.9033
No log 6.2069 180 0.7641 0.7421 0.7641 0.8741
No log 6.2759 182 0.8419 0.7310 0.8419 0.9176
No log 6.3448 184 0.8824 0.6849 0.8824 0.9394
No log 6.4138 186 0.9632 0.6746 0.9632 0.9814
No log 6.4828 188 1.3130 0.6111 1.3130 1.1458
No log 6.5517 190 1.5446 0.5381 1.5446 1.2428
No log 6.6207 192 1.4175 0.5341 1.4175 1.1906
No log 6.6897 194 1.2183 0.5269 1.2183 1.1038
No log 6.7586 196 0.9783 0.6667 0.9783 0.9891
No log 6.8276 198 0.8528 0.7123 0.8528 0.9235
No log 6.8966 200 0.8008 0.7083 0.8008 0.8949
No log 6.9655 202 0.7773 0.7260 0.7773 0.8817
No log 7.0345 204 0.8403 0.7284 0.8403 0.9167
No log 7.1034 206 0.9843 0.6905 0.9843 0.9921
No log 7.1724 208 0.9807 0.6341 0.9807 0.9903
No log 7.2414 210 0.9211 0.6104 0.9211 0.9598
No log 7.3103 212 0.9264 0.64 0.9264 0.9625
No log 7.3793 214 1.0600 0.5833 1.0600 1.0295
No log 7.4483 216 1.1221 0.5315 1.1221 1.0593
No log 7.5172 218 0.9922 0.6497 0.9922 0.9961
No log 7.5862 220 0.8326 0.7349 0.8326 0.9124
No log 7.6552 222 0.8553 0.7176 0.8553 0.9248
No log 7.7241 224 0.9785 0.7104 0.9785 0.9892
No log 7.7931 226 1.1088 0.7 1.1088 1.0530
No log 7.8621 228 0.7895 0.7143 0.7895 0.8885
No log 7.9310 230 0.7160 0.7882 0.7160 0.8461
No log 8.0 232 0.7336 0.7692 0.7336 0.8565
No log 8.0690 234 0.7866 0.6842 0.7866 0.8869
No log 8.1379 236 0.7921 0.6928 0.7921 0.8900
No log 8.2069 238 0.7985 0.6928 0.7985 0.8936
No log 8.2759 240 0.7896 0.6667 0.7896 0.8886
No log 8.3448 242 0.7917 0.6575 0.7917 0.8898
No log 8.4138 244 0.7872 0.7020 0.7872 0.8872
No log 8.4828 246 0.7911 0.7067 0.7911 0.8894
No log 8.5517 248 0.8106 0.6575 0.8106 0.9003
No log 8.6207 250 0.8308 0.6434 0.8308 0.9115
No log 8.6897 252 0.8075 0.6939 0.8075 0.8986
No log 8.7586 254 0.8676 0.7746 0.8676 0.9315
No log 8.8276 256 1.0132 0.6742 1.0132 1.0066
No log 8.8966 258 0.9511 0.7459 0.9511 0.9752
No log 8.9655 260 0.7777 0.8 0.7777 0.8819
No log 9.0345 262 0.7868 0.7125 0.7868 0.8870
No log 9.1034 264 0.8163 0.6839 0.8163 0.9035
No log 9.1724 266 0.7962 0.725 0.7962 0.8923
No log 9.2414 268 0.8449 0.7865 0.8449 0.9192
No log 9.3103 270 1.1005 0.6492 1.1005 1.0490
No log 9.3793 272 1.2615 0.6597 1.2615 1.1231
No log 9.4483 274 1.1648 0.6562 1.1648 1.0793
No log 9.5172 276 0.8886 0.7634 0.8886 0.9426
No log 9.5862 278 0.8145 0.7558 0.8145 0.9025
No log 9.6552 280 0.8484 0.6988 0.8484 0.9211
No log 9.7241 282 0.8585 0.7152 0.8585 0.9265
No log 9.7931 284 0.8771 0.6832 0.8771 0.9366
No log 9.8621 286 0.8573 0.7044 0.8573 0.9259
No log 9.9310 288 0.8200 0.7 0.8200 0.9055
No log 10.0 290 0.7948 0.6957 0.7948 0.8915
No log 10.0690 292 0.7605 0.7239 0.7605 0.8720
No log 10.1379 294 0.7445 0.8068 0.7445 0.8628
No log 10.2069 296 0.7970 0.7692 0.7970 0.8927
No log 10.2759 298 0.8836 0.7565 0.8836 0.9400
No log 10.3448 300 0.9003 0.7526 0.9003 0.9488
No log 10.4138 302 0.9169 0.7423 0.9169 0.9575
No log 10.4828 304 0.8101 0.7807 0.8101 0.9000
No log 10.5517 306 0.7533 0.7956 0.7533 0.8679
No log 10.6207 308 0.7353 0.7904 0.7353 0.8575
No log 10.6897 310 0.7700 0.6842 0.7700 0.8775
No log 10.7586 312 0.7911 0.6800 0.7911 0.8894
No log 10.8276 314 0.7746 0.7296 0.7746 0.8801
No log 10.8966 316 0.8168 0.8023 0.8168 0.9038
No log 10.9655 318 0.8922 0.7667 0.8922 0.9445
No log 11.0345 320 0.9599 0.7407 0.9599 0.9797
No log 11.1034 322 1.0171 0.7236 1.0171 1.0085
No log 11.1724 324 0.9528 0.7363 0.9528 0.9761
No log 11.2414 326 0.8644 0.7634 0.8644 0.9297
No log 11.3103 328 0.8576 0.7514 0.8576 0.9261
No log 11.3793 330 0.8801 0.7429 0.8801 0.9382
No log 11.4483 332 0.9340 0.6824 0.9340 0.9665
No log 11.5172 334 0.9134 0.6946 0.9134 0.9557
No log 11.5862 336 0.8656 0.7305 0.8656 0.9304
No log 11.6552 338 0.8310 0.75 0.8310 0.9116
No log 11.7241 340 0.8073 0.75 0.8073 0.8985
No log 11.7931 342 0.7575 0.7975 0.7575 0.8704
No log 11.8621 344 0.7427 0.7826 0.7427 0.8618
No log 11.9310 346 0.7473 0.75 0.7473 0.8645
No log 12.0 348 0.7407 0.7582 0.7407 0.8606
No log 12.0690 350 0.7327 0.7632 0.7327 0.8560
No log 12.1379 352 0.7554 0.7901 0.7554 0.8692
No log 12.2069 354 0.7738 0.7273 0.7738 0.8796
No log 12.2759 356 0.7198 0.7950 0.7198 0.8484
No log 12.3448 358 0.6938 0.7848 0.6938 0.8329
No log 12.4138 360 0.6848 0.7927 0.6848 0.8276
No log 12.4828 362 0.7134 0.7977 0.7134 0.8446
No log 12.5517 364 0.7954 0.7821 0.7954 0.8919
No log 12.6207 366 0.8997 0.7039 0.8997 0.9485
No log 12.6897 368 0.9580 0.7 0.9580 0.9788
No log 12.7586 370 0.8889 0.7222 0.8889 0.9428
No log 12.8276 372 0.7956 0.7865 0.7956 0.8920
No log 12.8966 374 0.7573 0.7886 0.7573 0.8702
No log 12.9655 376 0.7660 0.7421 0.7660 0.8752
No log 13.0345 378 0.7955 0.7879 0.7955 0.8919
No log 13.1034 380 0.8673 0.7674 0.8673 0.9313
No log 13.1724 382 0.9332 0.7024 0.9332 0.9660
No log 13.2414 384 1.0197 0.6590 1.0197 1.0098
No log 13.3103 386 1.0157 0.6554 1.0157 1.0078
No log 13.3793 388 0.8997 0.7356 0.8997 0.9485
No log 13.4483 390 0.7954 0.7746 0.7954 0.8919
No log 13.5172 392 0.7655 0.7547 0.7655 0.8749
No log 13.5862 394 0.7730 0.7613 0.7730 0.8792
No log 13.6552 396 0.8271 0.6853 0.8271 0.9094
No log 13.7241 398 0.9468 0.6569 0.9468 0.9730
No log 13.7931 400 0.9563 0.6569 0.9563 0.9779
No log 13.8621 402 0.8628 0.7007 0.8628 0.9289
No log 13.9310 404 0.7435 0.7467 0.7435 0.8623
No log 14.0 406 0.7537 0.7394 0.7537 0.8682
No log 14.0690 408 0.8230 0.7241 0.8230 0.9072
No log 14.1379 410 0.7786 0.7657 0.7786 0.8824
No log 14.2069 412 0.7055 0.7931 0.7055 0.8399
No log 14.2759 414 0.6882 0.7805 0.6882 0.8296
No log 14.3448 416 0.6927 0.7805 0.6927 0.8323
No log 14.4138 418 0.6995 0.7702 0.6995 0.8363
No log 14.4828 420 0.7123 0.775 0.7123 0.8440
No log 14.5517 422 0.7177 0.7532 0.7177 0.8472
No log 14.6207 424 0.7228 0.7582 0.7228 0.8502
No log 14.6897 426 0.7337 0.7248 0.7337 0.8566
No log 14.7586 428 0.7283 0.7248 0.7283 0.8534
No log 14.8276 430 0.7075 0.7532 0.7075 0.8411
No log 14.8966 432 0.6867 0.7662 0.6867 0.8287
No log 14.9655 434 0.6893 0.7826 0.6893 0.8302
No log 15.0345 436 0.7306 0.7702 0.7306 0.8548
No log 15.1034 438 0.8333 0.7296 0.8333 0.9129
No log 15.1724 440 0.8610 0.7215 0.8610 0.9279
No log 15.2414 442 0.7748 0.7453 0.7748 0.8802
No log 15.3103 444 0.6832 0.7805 0.6832 0.8265
No log 15.3793 446 0.6954 0.7484 0.6954 0.8339
No log 15.4483 448 0.7346 0.7134 0.7345 0.8571
No log 15.5172 450 0.7176 0.7515 0.7176 0.8471
No log 15.5862 452 0.7162 0.7907 0.7162 0.8463
No log 15.6552 454 0.7378 0.7778 0.7378 0.8590
No log 15.7241 456 0.7479 0.7598 0.7479 0.8648
No log 15.7931 458 0.7768 0.7760 0.7768 0.8814
No log 15.8621 460 0.7714 0.7802 0.7714 0.8783
No log 15.9310 462 0.7263 0.7907 0.7263 0.8522
No log 16.0 464 0.7163 0.7778 0.7163 0.8464
No log 16.0690 466 0.7068 0.7643 0.7068 0.8407
No log 16.1379 468 0.6784 0.7811 0.6784 0.8236
No log 16.2069 470 0.6607 0.7907 0.6607 0.8128
No log 16.2759 472 0.6877 0.7933 0.6877 0.8293
No log 16.3448 474 0.7051 0.7933 0.7051 0.8397
No log 16.4138 476 0.7009 0.7933 0.7009 0.8372
No log 16.4828 478 0.6825 0.8068 0.6825 0.8262
No log 16.5517 480 0.6706 0.8136 0.6706 0.8189
No log 16.6207 482 0.6713 0.7977 0.6713 0.8193
No log 16.6897 484 0.6641 0.7907 0.6641 0.8149
No log 16.7586 486 0.6894 0.7791 0.6894 0.8303
No log 16.8276 488 0.7152 0.7836 0.7152 0.8457
No log 16.8966 490 0.7523 0.7765 0.7523 0.8674
No log 16.9655 492 0.7576 0.7836 0.7576 0.8704
No log 17.0345 494 0.7664 0.7317 0.7664 0.8754
No log 17.1034 496 0.7625 0.7407 0.7625 0.8732
No log 17.1724 498 0.7633 0.7262 0.7633 0.8737
0.4137 17.2414 500 0.7720 0.7294 0.7720 0.8786
0.4137 17.3103 502 0.8039 0.7368 0.8039 0.8966
0.4137 17.3793 504 0.8071 0.7239 0.8071 0.8984
0.4137 17.4483 506 0.7632 0.7296 0.7632 0.8736
0.4137 17.5172 508 0.7324 0.7758 0.7324 0.8558
0.4137 17.5862 510 0.7312 0.7907 0.7312 0.8551
0.4137 17.6552 512 0.7596 0.7910 0.7596 0.8715
0.4137 17.7241 514 0.7765 0.7771 0.7765 0.8812
0.4137 17.7931 516 0.7413 0.8068 0.7413 0.8610
0.4137 17.8621 518 0.7126 0.7836 0.7126 0.8442
0.4137 17.9310 520 0.7109 0.7857 0.7109 0.8431
0.4137 18.0 522 0.7136 0.7952 0.7136 0.8447
0.4137 18.0690 524 0.7344 0.7368 0.7344 0.8570
0.4137 18.1379 526 0.7385 0.7368 0.7385 0.8594
0.4137 18.2069 528 0.7218 0.7875 0.7218 0.8496
0.4137 18.2759 530 0.7269 0.7952 0.7269 0.8526
0.4137 18.3448 532 0.7461 0.7929 0.7461 0.8637
0.4137 18.4138 534 0.7643 0.7170 0.7643 0.8743
0.4137 18.4828 536 0.7632 0.7170 0.7632 0.8736
0.4137 18.5517 538 0.7537 0.7368 0.7537 0.8682
0.4137 18.6207 540 0.7424 0.7248 0.7424 0.8616
0.4137 18.6897 542 0.7538 0.7517 0.7538 0.8682
0.4137 18.7586 544 0.7717 0.7297 0.7717 0.8785
0.4137 18.8276 546 0.8141 0.6712 0.8141 0.9023
0.4137 18.8966 548 0.8508 0.6803 0.8508 0.9224
0.4137 18.9655 550 0.8553 0.6803 0.8553 0.9248
0.4137 19.0345 552 0.8456 0.6892 0.8456 0.9196
0.4137 19.1034 554 0.8154 0.7226 0.8154 0.9030

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task1_organization

Finetuned
(4023)
this model