ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k10_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6707
  • Qwk: 0.7297
  • Mse: 0.6707
  • Rmse: 0.8190

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0417 2 6.9353 0.0242 6.9353 2.6335
No log 0.0833 4 4.4790 0.0772 4.4790 2.1164
No log 0.125 6 3.2888 0.0119 3.2888 1.8135
No log 0.1667 8 2.2506 0.1277 2.2506 1.5002
No log 0.2083 10 2.0866 0.2206 2.0866 1.4445
No log 0.25 12 2.5362 0.0530 2.5362 1.5926
No log 0.2917 14 3.0554 0.0732 3.0554 1.7480
No log 0.3333 16 2.7272 0.0892 2.7272 1.6514
No log 0.375 18 2.3546 0.12 2.3546 1.5345
No log 0.4167 20 1.8985 0.3307 1.8985 1.3779
No log 0.4583 22 1.7476 0.384 1.7476 1.3220
No log 0.5 24 1.6936 0.4186 1.6936 1.3014
No log 0.5417 26 2.0477 0.2740 2.0477 1.4310
No log 0.5833 28 2.6982 0.1618 2.6982 1.6426
No log 0.625 30 2.2991 0.2469 2.2991 1.5163
No log 0.6667 32 2.2540 0.2545 2.2540 1.5013
No log 0.7083 34 1.7207 0.3378 1.7207 1.3118
No log 0.75 36 1.7086 0.3378 1.7086 1.3071
No log 0.7917 38 1.9524 0.3046 1.9524 1.3973
No log 0.8333 40 2.2150 0.2112 2.2150 1.4883
No log 0.875 42 2.5473 0.2339 2.5473 1.5960
No log 0.9167 44 2.1471 0.2454 2.1471 1.4653
No log 0.9583 46 1.3195 0.5036 1.3195 1.1487
No log 1.0 48 1.0882 0.5634 1.0882 1.0432
No log 1.0417 50 1.2028 0.5949 1.2028 1.0967
No log 1.0833 52 1.1183 0.6329 1.1183 1.0575
No log 1.125 54 1.0062 0.6531 1.0062 1.0031
No log 1.1667 56 1.0892 0.6267 1.0892 1.0436
No log 1.2083 58 0.9252 0.6928 0.9252 0.9619
No log 1.25 60 0.7703 0.7432 0.7703 0.8776
No log 1.2917 62 0.7927 0.7162 0.7927 0.8903
No log 1.3333 64 1.0555 0.6797 1.0555 1.0274
No log 1.375 66 0.9985 0.7020 0.9985 0.9993
No log 1.4167 68 0.9014 0.6846 0.9014 0.9494
No log 1.4583 70 0.8775 0.7034 0.8775 0.9367
No log 1.5 72 0.9337 0.6712 0.9337 0.9663
No log 1.5417 74 1.0171 0.7018 1.0171 1.0085
No log 1.5833 76 0.9084 0.6977 0.9084 0.9531
No log 1.625 78 0.7359 0.7784 0.7359 0.8579
No log 1.6667 80 0.7414 0.7974 0.7414 0.8611
No log 1.7083 82 0.6767 0.7950 0.6767 0.8226
No log 1.75 84 0.6756 0.8049 0.6756 0.8220
No log 1.7917 86 0.6771 0.8166 0.6771 0.8228
No log 1.8333 88 0.6968 0.7763 0.6968 0.8347
No log 1.875 90 0.7148 0.7534 0.7148 0.8455
No log 1.9167 92 0.6773 0.8125 0.6773 0.8230
No log 1.9583 94 0.6679 0.8125 0.6679 0.8173
No log 2.0 96 0.7263 0.7821 0.7263 0.8522
No log 2.0417 98 0.9495 0.6590 0.9495 0.9744
No log 2.0833 100 1.1763 0.6444 1.1763 1.0846
No log 2.125 102 0.9287 0.6667 0.9287 0.9637
No log 2.1667 104 0.7799 0.75 0.7799 0.8831
No log 2.2083 106 0.7848 0.7383 0.7848 0.8859
No log 2.25 108 0.7864 0.7821 0.7864 0.8868
No log 2.2917 110 0.7825 0.7722 0.7825 0.8846
No log 2.3333 112 0.7870 0.7347 0.7870 0.8872
No log 2.375 114 0.7800 0.7324 0.7800 0.8832
No log 2.4167 116 0.8505 0.7101 0.8505 0.9222
No log 2.4583 118 0.7455 0.7429 0.7455 0.8634
No log 2.5 120 0.7355 0.7429 0.7355 0.8576
No log 2.5417 122 0.7288 0.7714 0.7288 0.8537
No log 2.5833 124 0.6769 0.7724 0.6769 0.8228
No log 2.625 126 0.6336 0.7895 0.6336 0.7960
No log 2.6667 128 0.6453 0.7673 0.6453 0.8033
No log 2.7083 130 0.7025 0.7778 0.7025 0.8381
No log 2.75 132 0.7077 0.7778 0.7077 0.8412
No log 2.7917 134 0.7092 0.7532 0.7092 0.8422
No log 2.8333 136 0.8005 0.7375 0.8005 0.8947
No log 2.875 138 0.8048 0.7375 0.8048 0.8971
No log 2.9167 140 0.8711 0.6962 0.8711 0.9333
No log 2.9583 142 0.9367 0.6832 0.9367 0.9678
No log 3.0 144 0.7746 0.7114 0.7746 0.8801
No log 3.0417 146 0.7565 0.7248 0.7565 0.8698
No log 3.0833 148 0.7896 0.75 0.7896 0.8886
No log 3.125 150 0.7726 0.7602 0.7726 0.8790
No log 3.1667 152 0.8252 0.7582 0.8252 0.9084
No log 3.2083 154 0.7344 0.7486 0.7344 0.8570
No log 3.25 156 0.7097 0.7619 0.7097 0.8425
No log 3.2917 158 0.7680 0.7473 0.7680 0.8764
No log 3.3333 160 1.0714 0.6961 1.0714 1.0351
No log 3.375 162 1.0565 0.7024 1.0565 1.0279
No log 3.4167 164 0.7881 0.7826 0.7881 0.8878
No log 3.4583 166 0.7905 0.7160 0.7905 0.8891
No log 3.5 168 0.7561 0.7561 0.7561 0.8695
No log 3.5417 170 0.7551 0.7977 0.7551 0.8689
No log 3.5833 172 0.9361 0.7079 0.9361 0.9675
No log 3.625 174 0.9440 0.72 0.9440 0.9716
No log 3.6667 176 0.7891 0.7578 0.7891 0.8883
No log 3.7083 178 0.7250 0.7922 0.7250 0.8515
No log 3.75 180 0.6964 0.7643 0.6964 0.8345
No log 3.7917 182 0.6862 0.7733 0.6862 0.8284
No log 3.8333 184 0.7255 0.7722 0.7255 0.8518
No log 3.875 186 0.8631 0.7453 0.8631 0.9290
No log 3.9167 188 1.0675 0.6705 1.0675 1.0332
No log 3.9583 190 1.0126 0.7006 1.0126 1.0063
No log 4.0 192 0.7375 0.7547 0.7375 0.8588
No log 4.0417 194 0.6918 0.75 0.6918 0.8318
No log 4.0833 196 0.6928 0.7429 0.6928 0.8323
No log 4.125 198 0.6373 0.7898 0.6373 0.7983
No log 4.1667 200 0.7546 0.7673 0.7546 0.8687
No log 4.2083 202 0.8838 0.7294 0.8838 0.9401
No log 4.25 204 0.8807 0.675 0.8807 0.9384
No log 4.2917 206 0.7339 0.7821 0.7339 0.8567
No log 4.3333 208 0.6734 0.7448 0.6734 0.8206
No log 4.375 210 0.6817 0.7465 0.6817 0.8256
No log 4.4167 212 0.6912 0.7172 0.6912 0.8314
No log 4.4583 214 0.7351 0.7613 0.7351 0.8574
No log 4.5 216 0.6486 0.7532 0.6486 0.8053
No log 4.5417 218 0.5883 0.8 0.5883 0.7670
No log 4.5833 220 0.5873 0.8 0.5873 0.7664
No log 4.625 222 0.6313 0.7711 0.6313 0.7946
No log 4.6667 224 0.7135 0.7956 0.7135 0.8447
No log 4.7083 226 0.7151 0.7607 0.7151 0.8456
No log 4.75 228 0.7410 0.8025 0.7410 0.8608
No log 4.7917 230 0.7212 0.7974 0.7212 0.8493
No log 4.8333 232 0.7417 0.7162 0.7417 0.8612
No log 4.875 234 0.7763 0.6849 0.7763 0.8811
No log 4.9167 236 0.7886 0.6939 0.7886 0.8881
No log 4.9583 238 0.8029 0.6968 0.8029 0.8960
No log 5.0 240 0.7283 0.7742 0.7283 0.8534
No log 5.0417 242 0.6637 0.7550 0.6637 0.8147
No log 5.0833 244 0.6983 0.7799 0.6983 0.8357
No log 5.125 246 0.7776 0.7771 0.7776 0.8818
No log 5.1667 248 0.6741 0.7485 0.6741 0.8210
No log 5.2083 250 0.5887 0.8075 0.5887 0.7673
No log 5.25 252 0.5899 0.7815 0.5899 0.7680
No log 5.2917 254 0.6457 0.7376 0.6457 0.8035
No log 5.3333 256 0.7194 0.7234 0.7194 0.8482
No log 5.375 258 0.7309 0.7143 0.7309 0.8549
No log 5.4167 260 0.7038 0.7571 0.7038 0.8390
No log 5.4583 262 0.6987 0.7606 0.6987 0.8359
No log 5.5 264 0.6472 0.7733 0.6472 0.8045
No log 5.5417 266 0.6065 0.7922 0.6065 0.7788
No log 5.5833 268 0.5800 0.7949 0.5800 0.7615
No log 5.625 270 0.5647 0.7949 0.5647 0.7515
No log 5.6667 272 0.5579 0.7975 0.5579 0.7469
No log 5.7083 274 0.5855 0.7904 0.5855 0.7652
No log 5.75 276 0.6282 0.8140 0.6282 0.7926
No log 5.7917 278 0.7962 0.7727 0.7962 0.8923
No log 5.8333 280 0.8135 0.7630 0.8135 0.9019
No log 5.875 282 0.7183 0.8140 0.7183 0.8475
No log 5.9167 284 0.6176 0.7682 0.6176 0.7859
No log 5.9583 286 0.6306 0.7651 0.6306 0.7941
No log 6.0 288 0.6645 0.7651 0.6645 0.8151
No log 6.0417 290 0.6181 0.7651 0.6181 0.7862
No log 6.0833 292 0.6052 0.7742 0.6052 0.7780
No log 6.125 294 0.7256 0.7907 0.7256 0.8518
No log 6.1667 296 0.9083 0.6857 0.9083 0.9531
No log 6.2083 298 0.9336 0.6743 0.9336 0.9662
No log 6.25 300 0.7572 0.7355 0.7572 0.8702
No log 6.2917 302 0.6401 0.7755 0.6401 0.8001
No log 6.3333 304 0.6433 0.7465 0.6433 0.8020
No log 6.375 306 0.6240 0.7586 0.6240 0.7899
No log 6.4167 308 0.6174 0.7595 0.6174 0.7857
No log 6.4583 310 0.7163 0.7857 0.7163 0.8463
No log 6.5 312 0.8312 0.7590 0.8312 0.9117
No log 6.5417 314 0.8339 0.7436 0.8339 0.9132
No log 6.5833 316 0.7929 0.7083 0.7929 0.8905
No log 6.625 318 0.7746 0.6861 0.7746 0.8801
No log 6.6667 320 0.7283 0.7397 0.7283 0.8534
No log 6.7083 322 0.7539 0.7397 0.7539 0.8682
No log 6.75 324 0.6673 0.7625 0.6673 0.8169
No log 6.7917 326 0.5540 0.8 0.5540 0.7443
No log 6.8333 328 0.5245 0.8077 0.5245 0.7243
No log 6.875 330 0.5949 0.7826 0.5949 0.7713
No log 6.9167 332 0.5957 0.7826 0.5957 0.7718
No log 6.9583 334 0.5626 0.8 0.5626 0.7501
No log 7.0 336 0.5994 0.7808 0.5994 0.7742
No log 7.0417 338 0.6393 0.7619 0.6393 0.7995
No log 7.0833 340 0.7149 0.7483 0.7149 0.8455
No log 7.125 342 0.7862 0.7211 0.7862 0.8867
No log 7.1667 344 0.8793 0.6538 0.8793 0.9377
No log 7.2083 346 0.8207 0.7170 0.8207 0.9059
No log 7.25 348 0.6936 0.7651 0.6936 0.8328
No log 7.2917 350 0.6212 0.7867 0.6212 0.7882
No log 7.3333 352 0.5867 0.7867 0.5867 0.7660
No log 7.375 354 0.5472 0.7947 0.5472 0.7398
No log 7.4167 356 0.5300 0.7867 0.5300 0.7280
No log 7.4583 358 0.5411 0.7919 0.5411 0.7356
No log 7.5 360 0.5503 0.7867 0.5503 0.7418
No log 7.5417 362 0.5701 0.7763 0.5701 0.7551
No log 7.5833 364 0.5659 0.7763 0.5659 0.7522
No log 7.625 366 0.6643 0.7607 0.6643 0.8150
No log 7.6667 368 0.6895 0.7711 0.6895 0.8304
No log 7.7083 370 0.6201 0.7692 0.6201 0.7874
No log 7.75 372 0.5650 0.7815 0.5650 0.7517
No log 7.7917 374 0.6058 0.7681 0.6058 0.7783
No log 7.8333 376 0.6504 0.7313 0.6504 0.8065
No log 7.875 378 0.6286 0.7660 0.6286 0.7928
No log 7.9167 380 0.6306 0.7550 0.6306 0.7941
No log 7.9583 382 0.6276 0.7792 0.6276 0.7922
No log 8.0 384 0.5998 0.7712 0.5998 0.7745
No log 8.0417 386 0.5956 0.7712 0.5956 0.7717
No log 8.0833 388 0.5986 0.7895 0.5986 0.7737
No log 8.125 390 0.6130 0.7867 0.6130 0.7830
No log 8.1667 392 0.6360 0.7534 0.6360 0.7975
No log 8.2083 394 0.7127 0.7361 0.7127 0.8442
No log 8.25 396 0.8023 0.6759 0.8023 0.8957
No log 8.2917 398 0.7536 0.7211 0.7536 0.8681
No log 8.3333 400 0.6456 0.7619 0.6456 0.8035
No log 8.375 402 0.5776 0.7947 0.5776 0.7600
No log 8.4167 404 0.5347 0.7733 0.5347 0.7312
No log 8.4583 406 0.5594 0.8105 0.5594 0.7479
No log 8.5 408 0.6244 0.7785 0.6244 0.7902
No log 8.5417 410 0.6126 0.8052 0.6126 0.7827
No log 8.5833 412 0.5754 0.8 0.5754 0.7586
No log 8.625 414 0.8664 0.7650 0.8664 0.9308
No log 8.6667 416 1.0743 0.6554 1.0743 1.0365
No log 8.7083 418 0.9483 0.7086 0.9483 0.9738
No log 8.75 420 0.7267 0.7436 0.7267 0.8525
No log 8.7917 422 0.6473 0.7871 0.6473 0.8045
No log 8.8333 424 0.6169 0.7949 0.6169 0.7854
No log 8.875 426 0.5954 0.7799 0.5954 0.7716
No log 8.9167 428 0.6172 0.7722 0.6172 0.7856
No log 8.9583 430 0.5988 0.7799 0.5988 0.7738
No log 9.0 432 0.5867 0.7843 0.5867 0.7659
No log 9.0417 434 0.5850 0.7947 0.5850 0.7649
No log 9.0833 436 0.5745 0.8 0.5745 0.7579
No log 9.125 438 0.6503 0.7482 0.6503 0.8064
No log 9.1667 440 0.7336 0.7338 0.7336 0.8565
No log 9.2083 442 0.6721 0.7586 0.6721 0.8198
No log 9.25 444 0.5816 0.7755 0.5816 0.7626
No log 9.2917 446 0.5588 0.8188 0.5588 0.7475
No log 9.3333 448 0.5643 0.8212 0.5643 0.7512
No log 9.375 450 0.5556 0.8108 0.5556 0.7454
No log 9.4167 452 0.5478 0.8158 0.5478 0.7401
No log 9.4583 454 0.5745 0.7826 0.5745 0.7580
No log 9.5 456 0.6593 0.7811 0.6593 0.8120
No log 9.5417 458 0.6710 0.7831 0.6710 0.8192
No log 9.5833 460 0.5805 0.775 0.5805 0.7619
No log 9.625 462 0.5308 0.8129 0.5308 0.7286
No log 9.6667 464 0.5526 0.7815 0.5526 0.7434
No log 9.7083 466 0.5768 0.7815 0.5768 0.7595
No log 9.75 468 0.6466 0.7898 0.6466 0.8041
No log 9.7917 470 0.7214 0.7673 0.7214 0.8493
No log 9.8333 472 0.7455 0.75 0.7455 0.8634
No log 9.875 474 0.7362 0.7853 0.7362 0.8580
No log 9.9167 476 0.6786 0.7975 0.6786 0.8238
No log 9.9583 478 0.6340 0.8025 0.6340 0.7962
No log 10.0 480 0.6059 0.8101 0.6059 0.7784
No log 10.0417 482 0.6016 0.8205 0.6016 0.7756
No log 10.0833 484 0.6051 0.7922 0.6051 0.7779
No log 10.125 486 0.6213 0.7712 0.6213 0.7882
No log 10.1667 488 0.6390 0.8 0.6390 0.7994
No log 10.2083 490 0.6318 0.8025 0.6318 0.7949
No log 10.25 492 0.6005 0.7826 0.6005 0.7749
No log 10.2917 494 0.6222 0.7730 0.6222 0.7888
No log 10.3333 496 0.7344 0.7857 0.7344 0.8569
No log 10.375 498 0.7116 0.7758 0.7116 0.8436
0.335 10.4167 500 0.5833 0.7925 0.5833 0.7638
0.335 10.4583 502 0.5404 0.8205 0.5404 0.7351
0.335 10.5 504 0.5392 0.8280 0.5392 0.7343
0.335 10.5417 506 0.5602 0.8280 0.5602 0.7485
0.335 10.5833 508 0.5997 0.7975 0.5997 0.7744
0.335 10.625 510 0.6117 0.8075 0.6117 0.7821
0.335 10.6667 512 0.5812 0.7975 0.5812 0.7624
0.335 10.7083 514 0.5710 0.8176 0.5710 0.7556
0.335 10.75 516 0.5892 0.8272 0.5892 0.7676
0.335 10.7917 518 0.6422 0.8075 0.6422 0.8013
0.335 10.8333 520 0.7069 0.7875 0.7069 0.8408
0.335 10.875 522 0.7057 0.7875 0.7057 0.8401
0.335 10.9167 524 0.6418 0.8125 0.6418 0.8011
0.335 10.9583 526 0.5933 0.8199 0.5933 0.7703
0.335 11.0 528 0.5609 0.8272 0.5609 0.7489
0.335 11.0417 530 0.5565 0.8199 0.5565 0.7460
0.335 11.0833 532 0.6033 0.8221 0.6033 0.7767
0.335 11.125 534 0.7523 0.7753 0.7523 0.8674
0.335 11.1667 536 0.8966 0.7104 0.8966 0.9469
0.335 11.2083 538 0.8856 0.6860 0.8856 0.9411
0.335 11.25 540 0.7651 0.7190 0.7651 0.8747
0.335 11.2917 542 0.6707 0.7297 0.6707 0.8190

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k10_task1_organization

Finetuned
(4023)
this model