ArabicNewSplits6_FineTuningAraBERTFreeze_run3_AugV5_k20_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8004
  • Qwk: 0.6209
  • Mse: 0.8004
  • Rmse: 0.8947

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.04 2 7.7835 -0.0439 7.7835 2.7899
No log 0.08 4 5.4899 -0.0378 5.4899 2.3430
No log 0.12 6 3.9661 0.0347 3.9661 1.9915
No log 0.16 8 3.0621 0.0785 3.0621 1.7499
No log 0.2 10 2.3236 0.1589 2.3236 1.5243
No log 0.24 12 1.7914 0.1661 1.7914 1.3384
No log 0.28 14 1.5119 0.2197 1.5119 1.2296
No log 0.32 16 1.3569 0.2292 1.3569 1.1649
No log 0.36 18 1.2172 0.2437 1.2172 1.1033
No log 0.4 20 1.1392 0.3738 1.1392 1.0673
No log 0.44 22 1.1035 0.4554 1.1035 1.0505
No log 0.48 24 1.0502 0.4742 1.0502 1.0248
No log 0.52 26 1.0055 0.4278 1.0055 1.0028
No log 0.56 28 0.9746 0.4181 0.9746 0.9872
No log 0.6 30 0.9426 0.4098 0.9426 0.9709
No log 0.64 32 0.9237 0.4804 0.9237 0.9611
No log 0.68 34 0.9463 0.5104 0.9463 0.9728
No log 0.72 36 1.0249 0.4475 1.0249 1.0124
No log 0.76 38 1.0430 0.4442 1.0430 1.0213
No log 0.8 40 1.0756 0.4201 1.0756 1.0371
No log 0.84 42 1.0380 0.3988 1.0380 1.0188
No log 0.88 44 0.9265 0.5071 0.9265 0.9626
No log 0.92 46 0.8933 0.5425 0.8933 0.9452
No log 0.96 48 0.9163 0.4977 0.9163 0.9573
No log 1.0 50 0.9743 0.4823 0.9743 0.9871
No log 1.04 52 1.0050 0.4572 1.0050 1.0025
No log 1.08 54 1.0302 0.4677 1.0302 1.0150
No log 1.12 56 1.0547 0.4355 1.0547 1.0270
No log 1.16 58 1.0129 0.4692 1.0129 1.0064
No log 1.2 60 0.9160 0.5129 0.9160 0.9571
No log 1.24 62 0.8892 0.5209 0.8892 0.9430
No log 1.28 64 0.9194 0.5174 0.9194 0.9589
No log 1.32 66 0.9611 0.5010 0.9611 0.9803
No log 1.3600 68 0.9343 0.5387 0.9343 0.9666
No log 1.4 70 0.9232 0.4960 0.9232 0.9608
No log 1.44 72 0.8665 0.5550 0.8665 0.9308
No log 1.48 74 0.8194 0.5862 0.8194 0.9052
No log 1.52 76 0.8685 0.5507 0.8685 0.9319
No log 1.56 78 0.9986 0.4810 0.9986 0.9993
No log 1.6 80 1.0296 0.4959 1.0296 1.0147
No log 1.6400 82 0.8786 0.5749 0.8786 0.9374
No log 1.6800 84 0.7600 0.6258 0.7600 0.8718
No log 1.72 86 0.7241 0.6431 0.7241 0.8509
No log 1.76 88 0.7538 0.6422 0.7538 0.8682
No log 1.8 90 0.8925 0.5710 0.8925 0.9447
No log 1.8400 92 0.9666 0.5316 0.9666 0.9832
No log 1.88 94 0.8858 0.5847 0.8858 0.9412
No log 1.92 96 0.7835 0.6586 0.7835 0.8851
No log 1.96 98 0.7156 0.6712 0.7156 0.8460
No log 2.0 100 0.7121 0.5982 0.7121 0.8439
No log 2.04 102 0.7272 0.5755 0.7272 0.8527
No log 2.08 104 0.7491 0.5912 0.7491 0.8655
No log 2.12 106 0.7750 0.6301 0.7750 0.8803
No log 2.16 108 0.7630 0.6462 0.7630 0.8735
No log 2.2 110 0.8057 0.6266 0.8057 0.8976
No log 2.24 112 0.8083 0.6037 0.8083 0.8990
No log 2.2800 114 0.8263 0.6084 0.8263 0.9090
No log 2.32 116 0.8804 0.5940 0.8804 0.9383
No log 2.36 118 0.8906 0.5861 0.8906 0.9437
No log 2.4 120 0.7961 0.6085 0.7961 0.8922
No log 2.44 122 0.7005 0.6565 0.7005 0.8370
No log 2.48 124 0.6847 0.6984 0.6847 0.8275
No log 2.52 126 0.6897 0.6955 0.6897 0.8305
No log 2.56 128 0.7113 0.6729 0.7113 0.8434
No log 2.6 130 0.7563 0.6682 0.7563 0.8697
No log 2.64 132 0.8745 0.6025 0.8745 0.9352
No log 2.68 134 0.9806 0.5375 0.9806 0.9903
No log 2.7200 136 0.9729 0.5614 0.9729 0.9864
No log 2.76 138 0.8397 0.6142 0.8397 0.9164
No log 2.8 140 0.7597 0.6239 0.7597 0.8716
No log 2.84 142 0.7326 0.6137 0.7326 0.8559
No log 2.88 144 0.7428 0.5739 0.7428 0.8619
No log 2.92 146 0.7455 0.6318 0.7455 0.8634
No log 2.96 148 0.8160 0.5994 0.8160 0.9033
No log 3.0 150 0.8186 0.6221 0.8186 0.9047
No log 3.04 152 0.8305 0.6042 0.8305 0.9113
No log 3.08 154 0.8393 0.5812 0.8393 0.9161
No log 3.12 156 0.8951 0.5586 0.8951 0.9461
No log 3.16 158 0.8426 0.5770 0.8426 0.9179
No log 3.2 160 0.7380 0.6750 0.7380 0.8591
No log 3.24 162 0.7142 0.6773 0.7142 0.8451
No log 3.2800 164 0.7239 0.6653 0.7239 0.8508
No log 3.32 166 0.8201 0.5958 0.8201 0.9056
No log 3.36 168 0.9990 0.5382 0.9990 0.9995
No log 3.4 170 1.0172 0.5300 1.0172 1.0086
No log 3.44 172 0.8754 0.5879 0.8754 0.9356
No log 3.48 174 0.7410 0.6294 0.7410 0.8608
No log 3.52 176 0.7561 0.5720 0.7561 0.8695
No log 3.56 178 0.7744 0.5469 0.7744 0.8800
No log 3.6 180 0.7209 0.6155 0.7209 0.8491
No log 3.64 182 0.7061 0.6513 0.7061 0.8403
No log 3.68 184 0.7728 0.6450 0.7728 0.8791
No log 3.7200 186 0.8442 0.6506 0.8442 0.9188
No log 3.76 188 0.7920 0.6352 0.7920 0.8899
No log 3.8 190 0.7024 0.6639 0.7024 0.8381
No log 3.84 192 0.6782 0.6760 0.6782 0.8235
No log 3.88 194 0.6694 0.6813 0.6694 0.8182
No log 3.92 196 0.6749 0.6893 0.6749 0.8215
No log 3.96 198 0.7209 0.6903 0.7209 0.8490
No log 4.0 200 0.7255 0.6903 0.7255 0.8518
No log 4.04 202 0.7188 0.7030 0.7188 0.8478
No log 4.08 204 0.7017 0.6887 0.7017 0.8377
No log 4.12 206 0.7132 0.6766 0.7132 0.8445
No log 4.16 208 0.7398 0.6400 0.7398 0.8601
No log 4.2 210 0.7129 0.6587 0.7129 0.8444
No log 4.24 212 0.6824 0.6528 0.6824 0.8261
No log 4.28 214 0.6667 0.6630 0.6667 0.8165
No log 4.32 216 0.6782 0.6558 0.6782 0.8236
No log 4.36 218 0.6935 0.6563 0.6935 0.8328
No log 4.4 220 0.6862 0.6479 0.6862 0.8284
No log 4.44 222 0.6800 0.6905 0.6800 0.8246
No log 4.48 224 0.7624 0.6280 0.7624 0.8731
No log 4.52 226 0.7897 0.6204 0.7897 0.8886
No log 4.5600 228 0.7385 0.6837 0.7385 0.8593
No log 4.6 230 0.7107 0.6829 0.7107 0.8430
No log 4.64 232 0.6995 0.6779 0.6995 0.8364
No log 4.68 234 0.7093 0.6687 0.7093 0.8422
No log 4.72 236 0.7176 0.6771 0.7176 0.8471
No log 4.76 238 0.7379 0.6696 0.7379 0.8590
No log 4.8 240 0.7745 0.6300 0.7745 0.8800
No log 4.84 242 0.7845 0.6202 0.7845 0.8857
No log 4.88 244 0.7620 0.6372 0.7620 0.8729
No log 4.92 246 0.7381 0.6462 0.7381 0.8591
No log 4.96 248 0.7332 0.6662 0.7332 0.8563
No log 5.0 250 0.7364 0.6591 0.7364 0.8581
No log 5.04 252 0.7356 0.6536 0.7356 0.8577
No log 5.08 254 0.8057 0.6399 0.8057 0.8976
No log 5.12 256 0.8358 0.6211 0.8358 0.9142
No log 5.16 258 0.7712 0.6636 0.7712 0.8782
No log 5.2 260 0.7345 0.6918 0.7345 0.8570
No log 5.24 262 0.7503 0.6441 0.7503 0.8662
No log 5.28 264 0.7439 0.6682 0.7439 0.8625
No log 5.32 266 0.7451 0.6630 0.7451 0.8632
No log 5.36 268 0.8225 0.6482 0.8225 0.9069
No log 5.4 270 0.8339 0.6044 0.8339 0.9132
No log 5.44 272 0.7596 0.6616 0.7596 0.8715
No log 5.48 274 0.7297 0.6676 0.7297 0.8543
No log 5.52 276 0.7188 0.6795 0.7188 0.8478
No log 5.5600 278 0.7247 0.6676 0.7247 0.8513
No log 5.6 280 0.7927 0.6219 0.7927 0.8904
No log 5.64 282 0.8282 0.6059 0.8282 0.9101
No log 5.68 284 0.7689 0.6288 0.7689 0.8769
No log 5.72 286 0.6811 0.7132 0.6811 0.8253
No log 5.76 288 0.7484 0.6396 0.7484 0.8651
No log 5.8 290 0.8623 0.6056 0.8623 0.9286
No log 5.84 292 0.8468 0.6170 0.8468 0.9202
No log 5.88 294 0.7575 0.6545 0.7575 0.8704
No log 5.92 296 0.7146 0.7208 0.7146 0.8453
No log 5.96 298 0.7767 0.6530 0.7767 0.8813
No log 6.0 300 0.7778 0.6509 0.7778 0.8819
No log 6.04 302 0.7241 0.6751 0.7241 0.8509
No log 6.08 304 0.7233 0.6795 0.7233 0.8504
No log 6.12 306 0.7711 0.6200 0.7711 0.8781
No log 6.16 308 0.7529 0.6308 0.7529 0.8677
No log 6.2 310 0.7076 0.6993 0.7076 0.8412
No log 6.24 312 0.6995 0.7233 0.6995 0.8363
No log 6.28 314 0.7010 0.7096 0.7010 0.8373
No log 6.32 316 0.7132 0.7185 0.7132 0.8445
No log 6.36 318 0.7033 0.7047 0.7033 0.8386
No log 6.4 320 0.7021 0.7077 0.7021 0.8379
No log 6.44 322 0.7151 0.6503 0.7151 0.8456
No log 6.48 324 0.7361 0.6304 0.7361 0.8580
No log 6.52 326 0.7390 0.6420 0.7390 0.8596
No log 6.5600 328 0.7488 0.6471 0.7488 0.8653
No log 6.6 330 0.7552 0.6323 0.7552 0.8691
No log 6.64 332 0.7578 0.6379 0.7578 0.8705
No log 6.68 334 0.7949 0.6413 0.7949 0.8916
No log 6.72 336 0.7804 0.6333 0.7804 0.8834
No log 6.76 338 0.7477 0.6635 0.7477 0.8647
No log 6.8 340 0.7703 0.6230 0.7703 0.8777
No log 6.84 342 0.7705 0.6285 0.7705 0.8778
No log 6.88 344 0.7483 0.6450 0.7483 0.8650
No log 6.92 346 0.7583 0.6657 0.7583 0.8708
No log 6.96 348 0.7793 0.6442 0.7793 0.8828
No log 7.0 350 0.7508 0.6427 0.7508 0.8665
No log 7.04 352 0.7337 0.6435 0.7337 0.8566
No log 7.08 354 0.7254 0.6485 0.7254 0.8517
No log 7.12 356 0.7254 0.6593 0.7254 0.8517
No log 7.16 358 0.7191 0.6593 0.7191 0.8480
No log 7.2 360 0.7048 0.6769 0.7048 0.8395
No log 7.24 362 0.7133 0.6820 0.7133 0.8445
No log 7.28 364 0.7159 0.6820 0.7159 0.8461
No log 7.32 366 0.7175 0.6911 0.7175 0.8471
No log 7.36 368 0.7324 0.6507 0.7324 0.8558
No log 7.4 370 0.7414 0.6772 0.7414 0.8611
No log 7.44 372 0.7600 0.6589 0.7600 0.8718
No log 7.48 374 0.8048 0.6333 0.8048 0.8971
No log 7.52 376 0.8314 0.6181 0.8314 0.9118
No log 7.5600 378 0.7927 0.6489 0.7927 0.8903
No log 7.6 380 0.7462 0.6927 0.7462 0.8638
No log 7.64 382 0.7324 0.6806 0.7324 0.8558
No log 7.68 384 0.7287 0.6735 0.7287 0.8536
No log 7.72 386 0.7355 0.6770 0.7355 0.8576
No log 7.76 388 0.7383 0.6770 0.7383 0.8592
No log 7.8 390 0.7224 0.6648 0.7224 0.8500
No log 7.84 392 0.7319 0.6129 0.7319 0.8555
No log 7.88 394 0.7600 0.6209 0.7600 0.8718
No log 7.92 396 0.7595 0.6692 0.7595 0.8715
No log 7.96 398 0.7377 0.6605 0.7377 0.8589
No log 8.0 400 0.7249 0.6968 0.7249 0.8514
No log 8.04 402 0.7296 0.6819 0.7296 0.8541
No log 8.08 404 0.7297 0.6747 0.7297 0.8542
No log 8.12 406 0.7138 0.6716 0.7138 0.8449
No log 8.16 408 0.7132 0.6511 0.7132 0.8445
No log 8.2 410 0.7133 0.6581 0.7133 0.8446
No log 8.24 412 0.7190 0.6679 0.7190 0.8479
No log 8.28 414 0.7212 0.6468 0.7212 0.8492
No log 8.32 416 0.7316 0.6685 0.7316 0.8553
No log 8.36 418 0.7674 0.6399 0.7674 0.8760
No log 8.4 420 0.7558 0.6798 0.7558 0.8694
No log 8.44 422 0.7366 0.6159 0.7366 0.8583
No log 8.48 424 0.7588 0.6206 0.7588 0.8711
No log 8.52 426 0.7873 0.6163 0.7873 0.8873
No log 8.56 428 0.7705 0.6392 0.7705 0.8778
No log 8.6 430 0.7455 0.6590 0.7455 0.8634
No log 8.64 432 0.7604 0.6711 0.7604 0.8720
No log 8.68 434 0.7558 0.6540 0.7558 0.8694
No log 8.72 436 0.7549 0.6218 0.7549 0.8688
No log 8.76 438 0.7614 0.6286 0.7614 0.8726
No log 8.8 440 0.7537 0.6341 0.7537 0.8681
No log 8.84 442 0.7861 0.6817 0.7861 0.8866
No log 8.88 444 0.8757 0.6041 0.8757 0.9358
No log 8.92 446 0.8589 0.6185 0.8589 0.9268
No log 8.96 448 0.8090 0.6340 0.8090 0.8994
No log 9.0 450 0.7513 0.6647 0.7513 0.8668
No log 9.04 452 0.7456 0.6301 0.7456 0.8635
No log 9.08 454 0.7410 0.6245 0.7410 0.8608
No log 9.12 456 0.7355 0.6291 0.7355 0.8576
No log 9.16 458 0.7303 0.6224 0.7303 0.8546
No log 9.2 460 0.7783 0.6316 0.7783 0.8822
No log 9.24 462 0.8273 0.6165 0.8273 0.9096
No log 9.28 464 0.7843 0.6316 0.7843 0.8856
No log 9.32 466 0.7259 0.6243 0.7259 0.8520
No log 9.36 468 0.7346 0.6587 0.7346 0.8571
No log 9.4 470 0.7503 0.6585 0.7503 0.8662
No log 9.44 472 0.7367 0.6720 0.7367 0.8583
No log 9.48 474 0.7208 0.6567 0.7208 0.8490
No log 9.52 476 0.7224 0.6831 0.7224 0.8499
No log 9.56 478 0.7192 0.6784 0.7192 0.8480
No log 9.6 480 0.7195 0.6603 0.7195 0.8482
No log 9.64 482 0.7193 0.6727 0.7193 0.8481
No log 9.68 484 0.7251 0.6764 0.7251 0.8515
No log 9.72 486 0.7452 0.6660 0.7452 0.8632
No log 9.76 488 0.7916 0.6190 0.7916 0.8897
No log 9.8 490 0.7823 0.6356 0.7823 0.8845
No log 9.84 492 0.7427 0.6799 0.7427 0.8618
No log 9.88 494 0.7411 0.6353 0.7411 0.8609
No log 9.92 496 0.7592 0.6648 0.7592 0.8713
No log 9.96 498 0.7464 0.6576 0.7464 0.8640
0.5613 10.0 500 0.7221 0.6382 0.7221 0.8498
0.5613 10.04 502 0.7575 0.6339 0.7575 0.8703
0.5613 10.08 504 0.7966 0.6230 0.7966 0.8925
0.5613 10.12 506 0.7750 0.6314 0.7750 0.8804
0.5613 10.16 508 0.7316 0.6252 0.7316 0.8553
0.5613 10.2 510 0.7295 0.6533 0.7295 0.8541
0.5613 10.24 512 0.7431 0.6572 0.7431 0.8620
0.5613 10.28 514 0.7613 0.6504 0.7613 0.8725
0.5613 10.32 516 0.7890 0.6424 0.7890 0.8883
0.5613 10.36 518 0.7872 0.6481 0.7872 0.8872
0.5613 10.4 520 0.8007 0.6263 0.8007 0.8948
0.5613 10.44 522 0.8097 0.6158 0.8097 0.8998
0.5613 10.48 524 0.7922 0.6257 0.7922 0.8900
0.5613 10.52 526 0.7860 0.6335 0.7860 0.8866
0.5613 10.56 528 0.7883 0.6422 0.7883 0.8879
0.5613 10.6 530 0.7650 0.6342 0.7650 0.8747
0.5613 10.64 532 0.7539 0.6511 0.7539 0.8683
0.5613 10.68 534 0.7442 0.6503 0.7442 0.8627
0.5613 10.72 536 0.7366 0.6422 0.7366 0.8582
0.5613 10.76 538 0.7275 0.6453 0.7275 0.8529
0.5613 10.8 540 0.7274 0.6543 0.7274 0.8529
0.5613 10.84 542 0.7367 0.6473 0.7367 0.8583
0.5613 10.88 544 0.7460 0.6740 0.7460 0.8637
0.5613 10.92 546 0.7550 0.6567 0.7550 0.8689
0.5613 10.96 548 0.7662 0.6527 0.7662 0.8754
0.5613 11.0 550 0.7604 0.6632 0.7604 0.8720
0.5613 11.04 552 0.7583 0.6453 0.7583 0.8708
0.5613 11.08 554 0.7576 0.6464 0.7576 0.8704
0.5613 11.12 556 0.7632 0.6302 0.7632 0.8736
0.5613 11.16 558 0.7749 0.6557 0.7749 0.8803
0.5613 11.2 560 0.7748 0.6434 0.7748 0.8802
0.5613 11.24 562 0.7824 0.6302 0.7824 0.8845
0.5613 11.28 564 0.7892 0.6362 0.7892 0.8884
0.5613 11.32 566 0.8024 0.6071 0.8024 0.8958
0.5613 11.36 568 0.8028 0.6177 0.8028 0.8960
0.5613 11.4 570 0.8004 0.6209 0.8004 0.8947

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run3_AugV5_k20_task1_organization

Finetuned
(4024)
this model