ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k10_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6081
  • Qwk: 0.7559
  • Mse: 0.6081
  • Rmse: 0.7798

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0345 2 5.2243 -0.0166 5.2243 2.2857
No log 0.0690 4 3.1087 0.0714 3.1087 1.7631
No log 0.1034 6 1.8929 0.1021 1.8929 1.3758
No log 0.1379 8 1.3782 0.1431 1.3782 1.1740
No log 0.1724 10 1.1824 0.2329 1.1824 1.0874
No log 0.2069 12 1.0824 0.2921 1.0824 1.0404
No log 0.2414 14 1.5608 0.1432 1.5608 1.2493
No log 0.2759 16 2.0265 0.1455 2.0265 1.4236
No log 0.3103 18 1.5728 0.0704 1.5728 1.2541
No log 0.3448 20 1.0859 0.2833 1.0859 1.0421
No log 0.3793 22 1.1702 0.1915 1.1702 1.0818
No log 0.4138 24 1.1548 0.2255 1.1548 1.0746
No log 0.4483 26 1.0662 0.2783 1.0662 1.0326
No log 0.4828 28 1.2632 0.2280 1.2632 1.1239
No log 0.5172 30 1.7071 0.1602 1.7071 1.3066
No log 0.5517 32 1.8158 0.1990 1.8158 1.3475
No log 0.5862 34 1.6274 0.2250 1.6274 1.2757
No log 0.6207 36 1.3798 0.3349 1.3798 1.1746
No log 0.6552 38 0.8657 0.5369 0.8657 0.9304
No log 0.6897 40 0.7015 0.6024 0.7015 0.8376
No log 0.7241 42 0.6688 0.6228 0.6688 0.8178
No log 0.7586 44 0.6787 0.6252 0.6787 0.8238
No log 0.7931 46 1.0722 0.4730 1.0722 1.0355
No log 0.8276 48 1.4155 0.4399 1.4155 1.1897
No log 0.8621 50 1.1514 0.4864 1.1514 1.0730
No log 0.8966 52 0.9581 0.6272 0.9581 0.9788
No log 0.9310 54 0.7696 0.6679 0.7696 0.8772
No log 0.9655 56 0.7401 0.6817 0.7401 0.8603
No log 1.0 58 1.0447 0.6072 1.0447 1.0221
No log 1.0345 60 1.2170 0.5331 1.2170 1.1032
No log 1.0690 62 0.9243 0.6367 0.9243 0.9614
No log 1.1034 64 0.6123 0.6862 0.6123 0.7825
No log 1.1379 66 0.7894 0.5308 0.7894 0.8885
No log 1.1724 68 0.9075 0.5275 0.9075 0.9526
No log 1.2069 70 0.7774 0.6089 0.7774 0.8817
No log 1.2414 72 0.7208 0.6496 0.7208 0.8490
No log 1.2759 74 0.7781 0.6785 0.7781 0.8821
No log 1.3103 76 0.7804 0.6943 0.7804 0.8834
No log 1.3448 78 0.7241 0.7331 0.7241 0.8509
No log 1.3793 80 0.6809 0.7502 0.6809 0.8252
No log 1.4138 82 0.7013 0.7273 0.7013 0.8374
No log 1.4483 84 0.8544 0.6572 0.8544 0.9243
No log 1.4828 86 0.8439 0.6331 0.8439 0.9186
No log 1.5172 88 0.6614 0.7292 0.6614 0.8132
No log 1.5517 90 0.5454 0.7740 0.5454 0.7385
No log 1.5862 92 0.6367 0.7350 0.6367 0.7979
No log 1.6207 94 0.6624 0.7455 0.6624 0.8139
No log 1.6552 96 0.5679 0.7390 0.5679 0.7536
No log 1.6897 98 0.5861 0.7180 0.5861 0.7656
No log 1.7241 100 0.6256 0.6944 0.6256 0.7910
No log 1.7586 102 0.5987 0.6969 0.5987 0.7738
No log 1.7931 104 0.6049 0.6923 0.6049 0.7777
No log 1.8276 106 0.5607 0.7238 0.5607 0.7488
No log 1.8621 108 0.5878 0.7248 0.5878 0.7667
No log 1.8966 110 0.5892 0.7497 0.5892 0.7676
No log 1.9310 112 0.7185 0.7368 0.7185 0.8476
No log 1.9655 114 0.9102 0.7075 0.9102 0.9540
No log 2.0 116 0.9119 0.7075 0.9119 0.9550
No log 2.0345 118 0.6893 0.7347 0.6893 0.8302
No log 2.0690 120 0.6244 0.7088 0.6244 0.7902
No log 2.1034 122 1.2291 0.4981 1.2291 1.1087
No log 2.1379 124 1.7068 0.4129 1.7068 1.3065
No log 2.1724 126 1.5772 0.4148 1.5772 1.2559
No log 2.2069 128 1.0533 0.5842 1.0533 1.0263
No log 2.2414 130 0.5995 0.7215 0.5995 0.7743
No log 2.2759 132 0.6592 0.7297 0.6592 0.8119
No log 2.3103 134 0.8314 0.6832 0.8314 0.9118
No log 2.3448 136 0.7702 0.6966 0.7702 0.8776
No log 2.3793 138 0.5875 0.7405 0.5875 0.7665
No log 2.4138 140 0.5430 0.7371 0.5430 0.7369
No log 2.4483 142 0.6261 0.6918 0.6261 0.7913
No log 2.4828 144 0.6544 0.6969 0.6544 0.8090
No log 2.5172 146 0.6384 0.7338 0.6384 0.7990
No log 2.5517 148 0.6426 0.7518 0.6426 0.8016
No log 2.5862 150 0.6601 0.7484 0.6601 0.8124
No log 2.6207 152 0.6925 0.7530 0.6925 0.8322
No log 2.6552 154 0.8640 0.6728 0.8640 0.9295
No log 2.6897 156 0.9724 0.6192 0.9724 0.9861
No log 2.7241 158 0.8159 0.6699 0.8159 0.9033
No log 2.7586 160 0.6983 0.7437 0.6983 0.8356
No log 2.7931 162 0.6496 0.7351 0.6496 0.8060
No log 2.8276 164 0.6247 0.7472 0.6247 0.7904
No log 2.8621 166 0.6155 0.7485 0.6155 0.7845
No log 2.8966 168 0.6136 0.748 0.6136 0.7833
No log 2.9310 170 0.6387 0.7278 0.6387 0.7992
No log 2.9655 172 0.7243 0.7031 0.7243 0.8511
No log 3.0 174 0.7327 0.7144 0.7327 0.8560
No log 3.0345 176 0.6867 0.7246 0.6867 0.8287
No log 3.0690 178 0.6778 0.7385 0.6778 0.8233
No log 3.1034 180 0.7370 0.7101 0.7370 0.8585
No log 3.1379 182 0.7151 0.7221 0.7151 0.8456
No log 3.1724 184 0.6599 0.7351 0.6599 0.8123
No log 3.2069 186 0.6284 0.7254 0.6284 0.7927
No log 3.2414 188 0.6142 0.7235 0.6142 0.7837
No log 3.2759 190 0.6117 0.7417 0.6117 0.7821
No log 3.3103 192 0.6034 0.7026 0.6034 0.7768
No log 3.3448 194 0.5813 0.7384 0.5813 0.7624
No log 3.3793 196 0.5880 0.7288 0.5880 0.7668
No log 3.4138 198 0.6218 0.7644 0.6218 0.7886
No log 3.4483 200 0.5987 0.7429 0.5987 0.7737
No log 3.4828 202 0.5930 0.7635 0.5930 0.7701
No log 3.5172 204 0.6533 0.7230 0.6533 0.8083
No log 3.5517 206 0.7108 0.6791 0.7108 0.8431
No log 3.5862 208 0.6654 0.7234 0.6654 0.8157
No log 3.6207 210 0.5798 0.7562 0.5798 0.7615
No log 3.6552 212 0.5629 0.7300 0.5629 0.7503
No log 3.6897 214 0.5788 0.7443 0.5788 0.7608
No log 3.7241 216 0.5808 0.7688 0.5808 0.7621
No log 3.7586 218 0.6276 0.7477 0.6276 0.7922
No log 3.7931 220 0.7562 0.6933 0.7562 0.8696
No log 3.8276 222 0.8794 0.6345 0.8794 0.9377
No log 3.8621 224 0.9613 0.5974 0.9613 0.9805
No log 3.8966 226 0.8304 0.6469 0.8304 0.9113
No log 3.9310 228 0.6593 0.7256 0.6593 0.8120
No log 3.9655 230 0.6093 0.7565 0.6093 0.7806
No log 4.0 232 0.6088 0.7316 0.6088 0.7803
No log 4.0345 234 0.6287 0.7342 0.6287 0.7929
No log 4.0690 236 0.6052 0.7279 0.6052 0.7779
No log 4.1034 238 0.5938 0.7236 0.5938 0.7706
No log 4.1379 240 0.6287 0.7085 0.6287 0.7929
No log 4.1724 242 0.6516 0.6905 0.6516 0.8072
No log 4.2069 244 0.6494 0.7166 0.6494 0.8059
No log 4.2414 246 0.6929 0.7297 0.6929 0.8324
No log 4.2759 248 0.7283 0.7340 0.7283 0.8534
No log 4.3103 250 0.7941 0.6893 0.7941 0.8911
No log 4.3448 252 0.7984 0.6932 0.7984 0.8935
No log 4.3793 254 0.7209 0.7387 0.7209 0.8490
No log 4.4138 256 0.6945 0.7627 0.6945 0.8334
No log 4.4483 258 0.7132 0.7477 0.7132 0.8445
No log 4.4828 260 0.6558 0.7566 0.6558 0.8098
No log 4.5172 262 0.6071 0.7531 0.6071 0.7792
No log 4.5517 264 0.5990 0.7330 0.5990 0.7740
No log 4.5862 266 0.6111 0.7224 0.6111 0.7817
No log 4.6207 268 0.6148 0.7164 0.6148 0.7841
No log 4.6552 270 0.6052 0.7399 0.6052 0.7779
No log 4.6897 272 0.6030 0.7282 0.6030 0.7765
No log 4.7241 274 0.6489 0.7533 0.6489 0.8055
No log 4.7586 276 0.6847 0.7510 0.6847 0.8275
No log 4.7931 278 0.6279 0.7603 0.6279 0.7924
No log 4.8276 280 0.6006 0.7557 0.6006 0.7750
No log 4.8621 282 0.5721 0.7382 0.5721 0.7564
No log 4.8966 284 0.5756 0.7498 0.5756 0.7587
No log 4.9310 286 0.5748 0.7497 0.5748 0.7581
No log 4.9655 288 0.5761 0.7497 0.5761 0.7590
No log 5.0 290 0.5783 0.7535 0.5783 0.7605
No log 5.0345 292 0.5790 0.7647 0.5790 0.7609
No log 5.0690 294 0.5916 0.7652 0.5916 0.7692
No log 5.1034 296 0.6041 0.7510 0.6041 0.7772
No log 5.1379 298 0.6218 0.7407 0.6218 0.7885
No log 5.1724 300 0.6493 0.7435 0.6493 0.8058
No log 5.2069 302 0.6544 0.7445 0.6544 0.8089
No log 5.2414 304 0.6490 0.7366 0.6490 0.8056
No log 5.2759 306 0.6289 0.7341 0.6289 0.7930
No log 5.3103 308 0.6178 0.7475 0.6178 0.7860
No log 5.3448 310 0.6282 0.7573 0.6282 0.7926
No log 5.3793 312 0.6307 0.7573 0.6307 0.7942
No log 5.4138 314 0.6511 0.7501 0.6511 0.8069
No log 5.4483 316 0.6431 0.7462 0.6431 0.8020
No log 5.4828 318 0.6698 0.7298 0.6698 0.8184
No log 5.5172 320 0.7333 0.7347 0.7333 0.8564
No log 5.5517 322 0.8413 0.7287 0.8413 0.9172
No log 5.5862 324 0.8786 0.7101 0.8786 0.9373
No log 5.6207 326 0.8836 0.7036 0.8836 0.9400
No log 5.6552 328 0.7906 0.7181 0.7906 0.8892
No log 5.6897 330 0.6970 0.7368 0.6970 0.8348
No log 5.7241 332 0.6517 0.7487 0.6517 0.8073
No log 5.7586 334 0.6567 0.7456 0.6567 0.8104
No log 5.7931 336 0.6905 0.7380 0.6905 0.8309
No log 5.8276 338 0.7219 0.7392 0.7219 0.8496
No log 5.8621 340 0.7636 0.7289 0.7636 0.8739
No log 5.8966 342 0.7902 0.7202 0.7902 0.8889
No log 5.9310 344 0.7788 0.7271 0.7788 0.8825
No log 5.9655 346 0.7184 0.7502 0.7184 0.8476
No log 6.0 348 0.6642 0.7517 0.6642 0.8150
No log 6.0345 350 0.6523 0.7501 0.6523 0.8076
No log 6.0690 352 0.6645 0.7435 0.6645 0.8152
No log 6.1034 354 0.6510 0.7442 0.6510 0.8069
No log 6.1379 356 0.6225 0.7320 0.6225 0.7890
No log 6.1724 358 0.6109 0.7420 0.6109 0.7816
No log 6.2069 360 0.6132 0.7474 0.6132 0.7831
No log 6.2414 362 0.6318 0.7531 0.6318 0.7949
No log 6.2759 364 0.6519 0.7243 0.6519 0.8074
No log 6.3103 366 0.7153 0.7577 0.7153 0.8457
No log 6.3448 368 0.7709 0.7359 0.7709 0.8780
No log 6.3793 370 0.7290 0.7638 0.7290 0.8538
No log 6.4138 372 0.6477 0.7384 0.6477 0.8048
No log 6.4483 374 0.5966 0.7353 0.5966 0.7724
No log 6.4828 376 0.5848 0.7652 0.5848 0.7647
No log 6.5172 378 0.5798 0.7596 0.5798 0.7614
No log 6.5517 380 0.5796 0.7548 0.5796 0.7613
No log 6.5862 382 0.5920 0.7452 0.5920 0.7694
No log 6.6207 384 0.6249 0.7653 0.6249 0.7905
No log 6.6552 386 0.6634 0.7595 0.6634 0.8145
No log 6.6897 388 0.6972 0.7362 0.6972 0.8350
No log 6.7241 390 0.7014 0.7362 0.7014 0.8375
No log 6.7586 392 0.6604 0.7647 0.6604 0.8127
No log 6.7931 394 0.5949 0.7567 0.5949 0.7713
No log 6.8276 396 0.5710 0.7458 0.5710 0.7557
No log 6.8621 398 0.5867 0.7380 0.5867 0.7659
No log 6.8966 400 0.5909 0.7326 0.5909 0.7687
No log 6.9310 402 0.5784 0.7564 0.5784 0.7605
No log 6.9655 404 0.5866 0.7558 0.5866 0.7659
No log 7.0 406 0.6498 0.7409 0.6498 0.8061
No log 7.0345 408 0.7300 0.7473 0.7300 0.8544
No log 7.0690 410 0.7481 0.7468 0.7481 0.8649
No log 7.1034 412 0.7055 0.7390 0.7055 0.8399
No log 7.1379 414 0.6391 0.7445 0.6391 0.7995
No log 7.1724 416 0.5939 0.7480 0.5939 0.7706
No log 7.2069 418 0.5974 0.7435 0.5974 0.7729
No log 7.2414 420 0.6096 0.7347 0.6096 0.7808
No log 7.2759 422 0.6083 0.7304 0.6083 0.7799
No log 7.3103 424 0.6004 0.7560 0.6004 0.7748
No log 7.3448 426 0.6074 0.7628 0.6074 0.7794
No log 7.3793 428 0.6174 0.7577 0.6174 0.7858
No log 7.4138 430 0.6254 0.7536 0.6254 0.7908
No log 7.4483 432 0.6273 0.7320 0.6273 0.7920
No log 7.4828 434 0.6369 0.7463 0.6369 0.7981
No log 7.5172 436 0.6384 0.7463 0.6384 0.7990
No log 7.5517 438 0.6298 0.7370 0.6298 0.7936
No log 7.5862 440 0.6335 0.7369 0.6335 0.7959
No log 7.6207 442 0.6313 0.7525 0.6313 0.7946
No log 7.6552 444 0.6339 0.7549 0.6339 0.7962
No log 7.6897 446 0.6308 0.7549 0.6308 0.7942
No log 7.7241 448 0.6103 0.7592 0.6103 0.7812
No log 7.7586 450 0.5952 0.7682 0.5952 0.7715
No log 7.7931 452 0.5828 0.7637 0.5828 0.7634
No log 7.8276 454 0.5778 0.7430 0.5778 0.7601
No log 7.8621 456 0.5795 0.7557 0.5795 0.7613
No log 7.8966 458 0.5857 0.7627 0.5857 0.7653
No log 7.9310 460 0.6048 0.7490 0.6048 0.7777
No log 7.9655 462 0.6208 0.7545 0.6208 0.7879
No log 8.0 464 0.6287 0.7582 0.6287 0.7929
No log 8.0345 466 0.6269 0.7565 0.6269 0.7918
No log 8.0690 468 0.6155 0.7549 0.6155 0.7845
No log 8.1034 470 0.5961 0.7548 0.5961 0.7721
No log 8.1379 472 0.5867 0.7679 0.5867 0.7660
No log 8.1724 474 0.5925 0.7630 0.5925 0.7697
No log 8.2069 476 0.5973 0.7576 0.5973 0.7728
No log 8.2414 478 0.5980 0.7756 0.5980 0.7733
No log 8.2759 480 0.5956 0.7643 0.5956 0.7717
No log 8.3103 482 0.5953 0.7596 0.5953 0.7716
No log 8.3448 484 0.5971 0.7533 0.5971 0.7727
No log 8.3793 486 0.6040 0.7459 0.6040 0.7772
No log 8.4138 488 0.6093 0.7550 0.6093 0.7806
No log 8.4483 490 0.6250 0.7509 0.6250 0.7905
No log 8.4828 492 0.6322 0.7509 0.6322 0.7951
No log 8.5172 494 0.6326 0.7509 0.6326 0.7953
No log 8.5517 496 0.6262 0.7509 0.6262 0.7913
No log 8.5862 498 0.6205 0.7531 0.6205 0.7877
0.4708 8.6207 500 0.6094 0.7576 0.6094 0.7806
0.4708 8.6552 502 0.6037 0.7576 0.6037 0.7769
0.4708 8.6897 504 0.6008 0.7520 0.6008 0.7751
0.4708 8.7241 506 0.5993 0.7520 0.5993 0.7742
0.4708 8.7586 508 0.5969 0.7562 0.5969 0.7726
0.4708 8.7931 510 0.5981 0.7562 0.5981 0.7733
0.4708 8.8276 512 0.5986 0.7520 0.5986 0.7737
0.4708 8.8621 514 0.5989 0.7520 0.5989 0.7739
0.4708 8.8966 516 0.5954 0.7520 0.5954 0.7716
0.4708 8.9310 518 0.5912 0.7520 0.5912 0.7689
0.4708 8.9655 520 0.5883 0.7562 0.5883 0.7670
0.4708 9.0 522 0.5846 0.7489 0.5846 0.7646
0.4708 9.0345 524 0.5811 0.7511 0.5811 0.7623
0.4708 9.0690 526 0.5797 0.7499 0.5797 0.7614
0.4708 9.1034 528 0.5793 0.7424 0.5793 0.7611
0.4708 9.1379 530 0.5802 0.7482 0.5802 0.7617
0.4708 9.1724 532 0.5831 0.7460 0.5831 0.7636
0.4708 9.2069 534 0.5867 0.7514 0.5867 0.7660
0.4708 9.2414 536 0.5905 0.7430 0.5905 0.7685
0.4708 9.2759 538 0.5943 0.7430 0.5943 0.7709
0.4708 9.3103 540 0.5972 0.7430 0.5972 0.7728
0.4708 9.3448 542 0.6005 0.7445 0.6005 0.7749
0.4708 9.3793 544 0.6051 0.7559 0.6051 0.7779
0.4708 9.4138 546 0.6112 0.7559 0.6112 0.7818
0.4708 9.4483 548 0.6165 0.7533 0.6165 0.7852
0.4708 9.4828 550 0.6186 0.7533 0.6186 0.7865
0.4708 9.5172 552 0.6188 0.7533 0.6188 0.7866
0.4708 9.5517 554 0.6187 0.7533 0.6187 0.7866
0.4708 9.5862 556 0.6184 0.7422 0.6184 0.7864
0.4708 9.6207 558 0.6189 0.7422 0.6189 0.7867
0.4708 9.6552 560 0.6181 0.7533 0.6181 0.7862
0.4708 9.6897 562 0.6172 0.7533 0.6172 0.7856
0.4708 9.7241 564 0.6163 0.7533 0.6163 0.7851
0.4708 9.7586 566 0.6155 0.7533 0.6155 0.7845
0.4708 9.7931 568 0.6132 0.7559 0.6132 0.7831
0.4708 9.8276 570 0.6114 0.7559 0.6114 0.7819
0.4708 9.8621 572 0.6102 0.7559 0.6102 0.7811
0.4708 9.8966 574 0.6094 0.7559 0.6094 0.7807
0.4708 9.9310 576 0.6087 0.7559 0.6087 0.7802
0.4708 9.9655 578 0.6082 0.7559 0.6082 0.7799
0.4708 10.0 580 0.6081 0.7559 0.6081 0.7798

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k10_task1_organization

Finetuned
(4023)
this model