ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k13_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8596
  • Qwk: 0.0277
  • Mse: 0.8596
  • Rmse: 0.9271

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 3.4981 0.0183 3.4981 1.8703
No log 0.1111 4 1.9415 0.0304 1.9415 1.3934
No log 0.1667 6 2.2890 -0.0129 2.2890 1.5129
No log 0.2222 8 1.0720 -0.1596 1.0720 1.0354
No log 0.2778 10 0.7585 0.1259 0.7585 0.8709
No log 0.3333 12 0.6890 0.0555 0.6890 0.8301
No log 0.3889 14 0.7427 0.1021 0.7427 0.8618
No log 0.4444 16 1.0700 -0.0468 1.0700 1.0344
No log 0.5 18 1.4275 0.0 1.4275 1.1948
No log 0.5556 20 1.3197 -0.0234 1.3197 1.1488
No log 0.6111 22 0.9273 -0.0638 0.9273 0.9630
No log 0.6667 24 0.7341 0.0334 0.7341 0.8568
No log 0.7222 26 0.6966 0.0964 0.6966 0.8346
No log 0.7778 28 0.7173 0.0374 0.7173 0.8469
No log 0.8333 30 0.8026 -0.0287 0.8026 0.8959
No log 0.8889 32 1.1066 -0.0234 1.1066 1.0519
No log 0.9444 34 1.2878 0.0 1.2878 1.1348
No log 1.0 36 1.1675 0.0 1.1675 1.0805
No log 1.0556 38 0.9497 -0.0385 0.9497 0.9745
No log 1.1111 40 0.8816 -0.0474 0.8816 0.9390
No log 1.1667 42 0.8042 0.0296 0.8042 0.8968
No log 1.2222 44 0.7306 0.0 0.7306 0.8548
No log 1.2778 46 0.7380 0.0 0.7380 0.8591
No log 1.3333 48 0.8916 -0.0033 0.8916 0.9443
No log 1.3889 50 1.6738 -0.0234 1.6738 1.2938
No log 1.4444 52 2.0510 -0.0014 2.0510 1.4321
No log 1.5 54 1.6192 -0.0234 1.6192 1.2725
No log 1.5556 56 1.1034 -0.0149 1.1034 1.0504
No log 1.6111 58 0.9690 0.0329 0.9690 0.9844
No log 1.6667 60 0.8986 0.0134 0.8986 0.9480
No log 1.7222 62 0.8575 0.0676 0.8575 0.9260
No log 1.7778 64 0.9086 0.0486 0.9086 0.9532
No log 1.8333 66 0.7974 0.0549 0.7974 0.8930
No log 1.8889 68 0.7076 -0.0035 0.7076 0.8412
No log 1.9444 70 0.7050 -0.0035 0.7050 0.8396
No log 2.0 72 0.7184 0.0 0.7184 0.8476
No log 2.0556 74 0.7349 -0.0033 0.7349 0.8573
No log 2.1111 76 0.7783 -0.0125 0.7783 0.8822
No log 2.1667 78 0.7863 -0.0204 0.7863 0.8867
No log 2.2222 80 0.9024 -0.0809 0.9024 0.9499
No log 2.2778 82 1.6615 -0.0235 1.6615 1.2890
No log 2.3333 84 2.2751 -0.0481 2.2751 1.5084
No log 2.3889 86 1.8943 -0.0247 1.8943 1.3763
No log 2.4444 88 1.2573 0.0196 1.2573 1.1213
No log 2.5 90 0.7951 -0.0675 0.7951 0.8917
No log 2.5556 92 0.7586 -0.1001 0.7586 0.8710
No log 2.6111 94 0.8129 -0.1001 0.8129 0.9016
No log 2.6667 96 0.7665 0.0099 0.7665 0.8755
No log 2.7222 98 0.7670 0.0191 0.7670 0.8758
No log 2.7778 100 0.7947 -0.0743 0.7947 0.8914
No log 2.8333 102 0.7786 0.1443 0.7786 0.8824
No log 2.8889 104 0.8029 0.0583 0.8029 0.8961
No log 2.9444 106 0.8611 0.1188 0.8611 0.9279
No log 3.0 108 0.7840 0.0308 0.7840 0.8854
No log 3.0556 110 0.7753 0.0757 0.7753 0.8805
No log 3.1111 112 0.8497 0.0647 0.8497 0.9218
No log 3.1667 114 0.8504 0.0592 0.8504 0.9222
No log 3.2222 116 0.8354 -0.0082 0.8354 0.9140
No log 3.2778 118 0.7808 0.0976 0.7808 0.8836
No log 3.3333 120 0.7635 0.0148 0.7635 0.8738
No log 3.3889 122 0.7070 0.0 0.7070 0.8408
No log 3.4444 124 0.9111 -0.1261 0.9111 0.9545
No log 3.5 126 1.0626 -0.0563 1.0626 1.0308
No log 3.5556 128 0.7295 0.1318 0.7295 0.8541
No log 3.6111 130 0.7290 -0.0520 0.7290 0.8538
No log 3.6667 132 0.8612 0.0776 0.8612 0.9280
No log 3.7222 134 0.8374 0.0297 0.8374 0.9151
No log 3.7778 136 0.7544 0.0814 0.7544 0.8685
No log 3.8333 138 1.0256 -0.0163 1.0256 1.0127
No log 3.8889 140 1.2539 0.0456 1.2539 1.1198
No log 3.9444 142 0.9353 -0.0118 0.9353 0.9671
No log 4.0 144 0.8308 0.1079 0.8308 0.9115
No log 4.0556 146 0.9827 0.0129 0.9827 0.9913
No log 4.1111 148 0.8661 0.0682 0.8661 0.9306
No log 4.1667 150 0.8248 -0.0274 0.8248 0.9082
No log 4.2222 152 0.9287 0.0277 0.9287 0.9637
No log 4.2778 154 0.8184 -0.0316 0.8184 0.9047
No log 4.3333 156 0.7857 0.1292 0.7857 0.8864
No log 4.3889 158 0.7794 0.0846 0.7794 0.8828
No log 4.4444 160 0.8355 0.0095 0.8355 0.9140
No log 4.5 162 0.8351 0.1431 0.8351 0.9138
No log 4.5556 164 0.7733 0.0179 0.7733 0.8794
No log 4.6111 166 0.8065 -0.0173 0.8065 0.8981
No log 4.6667 168 0.7321 -0.0451 0.7321 0.8556
No log 4.7222 170 0.6690 0.0436 0.6690 0.8179
No log 4.7778 172 0.7334 0.1817 0.7334 0.8564
No log 4.8333 174 0.7379 0.1387 0.7379 0.8590
No log 4.8889 176 0.7034 0.0524 0.7034 0.8387
No log 4.9444 178 0.7152 0.1081 0.7152 0.8457
No log 5.0 180 0.7575 0.1449 0.7575 0.8704
No log 5.0556 182 0.7778 0.1342 0.7778 0.8819
No log 5.1111 184 0.8571 0.0748 0.8571 0.9258
No log 5.1667 186 0.7952 0.1539 0.7952 0.8918
No log 5.2222 188 0.8359 0.0623 0.8359 0.9143
No log 5.2778 190 0.8076 0.1744 0.8076 0.8986
No log 5.3333 192 0.9731 -0.0157 0.9731 0.9865
No log 5.3889 194 1.0042 0.0157 1.0042 1.0021
No log 5.4444 196 0.8077 0.0040 0.8077 0.8987
No log 5.5 198 0.7779 0.2385 0.7779 0.8820
No log 5.5556 200 0.7358 0.1525 0.7358 0.8578
No log 5.6111 202 0.7342 0.0095 0.7342 0.8569
No log 5.6667 204 0.7142 0.1878 0.7142 0.8451
No log 5.7222 206 0.7385 0.1027 0.7385 0.8593
No log 5.7778 208 0.8400 -0.0111 0.8400 0.9165
No log 5.8333 210 0.8679 -0.0111 0.8679 0.9316
No log 5.8889 212 0.8442 0.1093 0.8442 0.9188
No log 5.9444 214 0.9809 -0.0076 0.9809 0.9904
No log 6.0 216 1.0527 -0.0163 1.0527 1.0260
No log 6.0556 218 0.8505 0.0409 0.8505 0.9222
No log 6.1111 220 0.7560 0.1362 0.7560 0.8695
No log 6.1667 222 0.8303 -0.0778 0.8303 0.9112
No log 6.2222 224 1.1494 0.0089 1.1494 1.0721
No log 6.2778 226 1.0498 -0.0194 1.0498 1.0246
No log 6.3333 228 1.0916 0.0046 1.0916 1.0448
No log 6.3889 230 1.0662 0.0067 1.0662 1.0325
No log 6.4444 232 0.8593 -0.0459 0.8593 0.9270
No log 6.5 234 0.7793 0.0570 0.7793 0.8828
No log 6.5556 236 0.8141 0.0285 0.8141 0.9023
No log 6.6111 238 0.7588 0.1644 0.7588 0.8711
No log 6.6667 240 0.9637 0.0182 0.9637 0.9817
No log 6.7222 242 0.8932 0.0618 0.8932 0.9451
No log 6.7778 244 0.7470 0.0081 0.7470 0.8643
No log 6.8333 246 0.8750 -0.0346 0.8750 0.9354
No log 6.8889 248 0.7966 0.0315 0.7966 0.8925
No log 6.9444 250 0.7563 0.2568 0.7563 0.8697
No log 7.0 252 0.9287 -0.0200 0.9287 0.9637
No log 7.0556 254 0.9313 -0.0182 0.9313 0.9650
No log 7.1111 256 0.7165 0.2754 0.7165 0.8464
No log 7.1667 258 0.7726 0.0361 0.7726 0.8790
No log 7.2222 260 0.8044 0.1159 0.8044 0.8969
No log 7.2778 262 0.7054 0.0949 0.7054 0.8399
No log 7.3333 264 0.6786 0.1282 0.6786 0.8237
No log 7.3889 266 0.7421 -0.0121 0.7421 0.8614
No log 7.4444 268 0.9023 0.0815 0.9023 0.9499
No log 7.5 270 0.9064 0.0815 0.9064 0.9521
No log 7.5556 272 0.7198 -0.0678 0.7198 0.8484
No log 7.6111 274 0.6932 0.1921 0.6932 0.8326
No log 7.6667 276 0.6945 0.2431 0.6945 0.8334
No log 7.7222 278 0.6926 0.2271 0.6926 0.8322
No log 7.7778 280 0.7191 -0.0366 0.7191 0.8480
No log 7.8333 282 0.7248 0.1146 0.7248 0.8513
No log 7.8889 284 0.9031 0.1150 0.9031 0.9503
No log 7.9444 286 0.8133 0.2054 0.8133 0.9018
No log 8.0 288 0.7267 0.2096 0.7267 0.8525
No log 8.0556 290 0.7496 -0.0257 0.7496 0.8658
No log 8.1111 292 0.7541 0.2009 0.7541 0.8684
No log 8.1667 294 0.8031 0.1758 0.8031 0.8961
No log 8.2222 296 0.7534 0.2078 0.7534 0.8680
No log 8.2778 298 0.7857 0.1553 0.7857 0.8864
No log 8.3333 300 0.8117 0.1440 0.8117 0.9010
No log 8.3889 302 0.8088 0.1440 0.8088 0.8993
No log 8.4444 304 0.7640 0.2914 0.7640 0.8741
No log 8.5 306 0.8075 0.2208 0.8075 0.8986
No log 8.5556 308 0.8686 0.0260 0.8686 0.9320
No log 8.6111 310 0.7726 0.2009 0.7726 0.8790
No log 8.6667 312 0.7651 0.2437 0.7651 0.8747
No log 8.7222 314 0.7597 0.0987 0.7597 0.8716
No log 8.7778 316 0.7796 0.1321 0.7796 0.8830
No log 8.8333 318 0.8107 0.1228 0.8107 0.9004
No log 8.8889 320 0.8362 0.3323 0.8362 0.9144
No log 8.9444 322 0.9294 -0.0143 0.9294 0.9641
No log 9.0 324 0.8837 -0.0143 0.8837 0.9401
No log 9.0556 326 0.7553 0.2138 0.7553 0.8691
No log 9.1111 328 0.7116 0.0964 0.7116 0.8435
No log 9.1667 330 0.7041 0.0964 0.7041 0.8391
No log 9.2222 332 0.7130 0.1902 0.7130 0.8444
No log 9.2778 334 0.7351 0.1254 0.7351 0.8574
No log 9.3333 336 0.7464 0.1612 0.7464 0.8639
No log 9.3889 338 0.8622 0.0609 0.8622 0.9285
No log 9.4444 340 0.8915 0.0182 0.8915 0.9442
No log 9.5 342 0.7655 0.2277 0.7655 0.8749
No log 9.5556 344 0.7269 0.1722 0.7269 0.8526
No log 9.6111 346 0.7458 0.1899 0.7458 0.8636
No log 9.6667 348 0.8715 0.0545 0.8715 0.9336
No log 9.7222 350 0.9162 0.0134 0.9162 0.9572
No log 9.7778 352 0.7724 0.2116 0.7724 0.8789
No log 9.8333 354 0.8127 -0.0781 0.8127 0.9015
No log 9.8889 356 0.8910 0.0129 0.8910 0.9439
No log 9.9444 358 0.7662 -0.0550 0.7662 0.8753
No log 10.0 360 0.7556 0.1449 0.7556 0.8692
No log 10.0556 362 0.7651 0.2141 0.7651 0.8747
No log 10.1111 364 0.7409 0.0269 0.7409 0.8608
No log 10.1667 366 0.8828 0.0673 0.8828 0.9396
No log 10.2222 368 1.0000 0.0492 1.0000 1.0000
No log 10.2778 370 0.8687 0.0301 0.8687 0.9320
No log 10.3333 372 0.7919 0.0611 0.7919 0.8899
No log 10.3889 374 0.8226 0.0456 0.8226 0.9070
No log 10.4444 376 0.7814 0.1146 0.7814 0.8839
No log 10.5 378 0.7606 0.0851 0.7606 0.8721
No log 10.5556 380 0.7755 0.0308 0.7755 0.8806
No log 10.6111 382 0.8208 0.2349 0.8208 0.9060
No log 10.6667 384 0.7887 0.1094 0.7887 0.8881
No log 10.7222 386 0.7947 0.0501 0.7947 0.8915
No log 10.7778 388 0.7979 0.0426 0.7979 0.8933
No log 10.8333 390 0.8804 -0.0079 0.8804 0.9383
No log 10.8889 392 0.9875 0.0182 0.9875 0.9937
No log 10.9444 394 0.8865 -0.0101 0.8865 0.9416
No log 11.0 396 0.8153 0.1508 0.8153 0.9029
No log 11.0556 398 0.8415 -0.0262 0.8415 0.9173
No log 11.1111 400 0.8800 -0.0033 0.8800 0.9381
No log 11.1667 402 0.8692 0.0409 0.8692 0.9323
No log 11.2222 404 0.8657 -0.0033 0.8657 0.9304
No log 11.2778 406 0.7804 0.0999 0.7804 0.8834
No log 11.3333 408 0.7726 0.0999 0.7726 0.8790
No log 11.3889 410 0.8638 0.0676 0.8638 0.9294
No log 11.4444 412 0.9351 -0.0558 0.9351 0.9670
No log 11.5 414 0.9927 -0.0236 0.9927 0.9963
No log 11.5556 416 0.8792 0.1107 0.8792 0.9376
No log 11.6111 418 0.7936 0.0559 0.7936 0.8908
No log 11.6667 420 0.9258 0.0092 0.9258 0.9622
No log 11.7222 422 0.8843 0.0092 0.8843 0.9404
No log 11.7778 424 0.7671 0.0148 0.7671 0.8758
No log 11.8333 426 0.7700 0.1047 0.7700 0.8775
No log 11.8889 428 0.8234 0.0316 0.8234 0.9074
No log 11.9444 430 0.8657 0.0233 0.8657 0.9304
No log 12.0 432 0.7799 0.0099 0.7799 0.8831
No log 12.0556 434 0.6961 0.0506 0.6961 0.8343
No log 12.1111 436 0.7261 0.0128 0.7261 0.8521
No log 12.1667 438 0.7306 0.0155 0.7306 0.8548
No log 12.2222 440 0.7395 0.0964 0.7395 0.8599
No log 12.2778 442 0.9057 0.0609 0.9057 0.9517
No log 12.3333 444 1.0162 -0.0236 1.0162 1.0081
No log 12.3889 446 0.9309 0.0609 0.9309 0.9648
No log 12.4444 448 0.8085 0.1431 0.8085 0.8992
No log 12.5 450 0.7670 0.1095 0.7670 0.8758
No log 12.5556 452 0.7435 0.0680 0.7435 0.8623
No log 12.6111 454 0.7421 0.1148 0.7421 0.8615
No log 12.6667 456 0.7319 0.1148 0.7319 0.8555
No log 12.7222 458 0.7736 0.2566 0.7736 0.8795
No log 12.7778 460 0.9088 0.0576 0.9088 0.9533
No log 12.8333 462 1.0832 -0.0316 1.0832 1.0408
No log 12.8889 464 1.0884 -0.0316 1.0884 1.0432
No log 12.9444 466 1.0166 -0.0301 1.0166 1.0083
No log 13.0 468 0.9025 -0.0097 0.9025 0.9500
No log 13.0556 470 0.8811 0.0277 0.8811 0.9387
No log 13.1111 472 0.9610 0.0111 0.9610 0.9803
No log 13.1667 474 0.8760 0.0277 0.8760 0.9359
No log 13.2222 476 0.8067 0.1800 0.8067 0.8982
No log 13.2778 478 0.7940 0.2169 0.7940 0.8911
No log 13.3333 480 0.8714 0.0207 0.8714 0.9335
No log 13.3889 482 0.9486 0.0026 0.9486 0.9740
No log 13.4444 484 0.9222 -0.0301 0.9222 0.9603
No log 13.5 486 0.8234 -0.0425 0.8234 0.9074
No log 13.5556 488 0.7526 0.0937 0.7526 0.8675
No log 13.6111 490 0.7699 0.0993 0.7699 0.8774
No log 13.6667 492 0.7457 0.0937 0.7457 0.8636
No log 13.7222 494 0.7628 0.1541 0.7628 0.8734
No log 13.7778 496 0.8338 0.1107 0.8338 0.9131
No log 13.8333 498 0.7758 0.1775 0.7758 0.8808
0.329 13.8889 500 0.7262 0.1287 0.7262 0.8522
0.329 13.9444 502 0.7668 0.1352 0.7668 0.8757
0.329 14.0 504 0.7549 0.0902 0.7549 0.8688
0.329 14.0556 506 0.7987 0.1485 0.7987 0.8937
0.329 14.1111 508 0.9521 0.0576 0.9521 0.9758
0.329 14.1667 510 0.9063 0.0233 0.9063 0.9520
0.329 14.2222 512 0.7772 0.1599 0.7772 0.8816
0.329 14.2778 514 0.7867 0.0291 0.7867 0.8870
0.329 14.3333 516 0.8192 -0.0093 0.8192 0.9051
0.329 14.3889 518 0.7614 -0.0214 0.7614 0.8726
0.329 14.4444 520 0.7070 0.1787 0.7070 0.8408
0.329 14.5 522 0.7734 0.1395 0.7734 0.8794
0.329 14.5556 524 0.8307 0.1453 0.8307 0.9114
0.329 14.6111 526 0.8212 0.1453 0.8212 0.9062
0.329 14.6667 528 0.7453 0.0376 0.7453 0.8633
0.329 14.7222 530 0.7548 -0.0121 0.7548 0.8688
0.329 14.7778 532 0.7466 -0.0724 0.7466 0.8641
0.329 14.8333 534 0.7640 0.1196 0.7640 0.8740
0.329 14.8889 536 0.7990 0.1758 0.7990 0.8939
0.329 14.9444 538 0.8761 0.0984 0.8761 0.9360
0.329 15.0 540 0.9018 0.0909 0.9018 0.9496
0.329 15.0556 542 0.8674 0.0642 0.8674 0.9313
0.329 15.1111 544 0.7751 0.2349 0.7751 0.8804
0.329 15.1667 546 0.7954 0.0424 0.7954 0.8919
0.329 15.2222 548 0.8382 0.0719 0.8382 0.9155
0.329 15.2778 550 0.8965 -0.0118 0.8965 0.9468
0.329 15.3333 552 0.9184 0.0182 0.9184 0.9583
0.329 15.3889 554 0.8596 0.0277 0.8596 0.9271

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k13_task3_organization

Finetuned
(4023)
this model