ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k14_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0235
  • Qwk: -0.0183
  • Mse: 1.0235
  • Rmse: 1.0117

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 3.6263 0.0048 3.6263 1.9043
No log 0.1053 4 1.9756 0.0704 1.9756 1.4056
No log 0.1579 6 1.2226 -0.0228 1.2226 1.1057
No log 0.2105 8 1.1342 -0.0193 1.1342 1.0650
No log 0.2632 10 0.8777 0.0404 0.8777 0.9369
No log 0.3158 12 0.7462 0.1259 0.7462 0.8639
No log 0.3684 14 0.7839 0.0191 0.7839 0.8854
No log 0.4211 16 0.9978 -0.0720 0.9978 0.9989
No log 0.4737 18 0.9106 -0.0949 0.9106 0.9542
No log 0.5263 20 0.7883 -0.0695 0.7883 0.8879
No log 0.5789 22 0.8355 -0.0408 0.8355 0.9141
No log 0.6316 24 0.8359 0.0512 0.8359 0.9143
No log 0.6842 26 0.7455 -0.0035 0.7455 0.8634
No log 0.7368 28 0.7782 -0.0215 0.7782 0.8822
No log 0.7895 30 0.8313 0.0129 0.8313 0.9118
No log 0.8421 32 0.7324 0.0 0.7324 0.8558
No log 0.8947 34 0.7628 -0.0679 0.7628 0.8734
No log 0.9474 36 0.9798 -0.1277 0.9798 0.9899
No log 1.0 38 1.5331 -0.0766 1.5331 1.2382
No log 1.0526 40 1.8055 0.0 1.8055 1.3437
No log 1.1053 42 1.5889 0.0 1.5889 1.2605
No log 1.1579 44 1.2874 -0.0247 1.2874 1.1346
No log 1.2105 46 0.9715 -0.0695 0.9715 0.9857
No log 1.2632 48 0.7724 -0.0131 0.7724 0.8789
No log 1.3158 50 0.7746 -0.1690 0.7746 0.8801
No log 1.3684 52 0.8297 0.0247 0.8297 0.9109
No log 1.4211 54 1.0151 -0.0518 1.0151 1.0075
No log 1.4737 56 1.1257 -0.0943 1.1257 1.0610
No log 1.5263 58 1.0346 -0.0845 1.0346 1.0172
No log 1.5789 60 0.9207 -0.0320 0.9207 0.9595
No log 1.6316 62 0.9557 0.1110 0.9557 0.9776
No log 1.6842 64 1.0016 0.0392 1.0016 1.0008
No log 1.7368 66 0.9567 -0.1449 0.9567 0.9781
No log 1.7895 68 1.0566 -0.1284 1.0566 1.0279
No log 1.8421 70 1.1914 -0.0221 1.1914 1.0915
No log 1.8947 72 1.3732 0.0 1.3732 1.1718
No log 1.9474 74 1.5385 0.0 1.5385 1.2404
No log 2.0 76 1.4376 0.0 1.4376 1.1990
No log 2.0526 78 1.1943 -0.0758 1.1943 1.0928
No log 2.1053 80 1.0463 -0.0992 1.0463 1.0229
No log 2.1579 82 0.9265 0.0065 0.9265 0.9625
No log 2.2105 84 0.9248 0.0146 0.9248 0.9616
No log 2.2632 86 0.9763 -0.0424 0.9763 0.9881
No log 2.3158 88 0.9926 -0.0500 0.9926 0.9963
No log 2.3684 90 0.8793 0.0099 0.8793 0.9377
No log 2.4211 92 0.7551 -0.0644 0.7551 0.8690
No log 2.4737 94 0.7362 0.0555 0.7362 0.8580
No log 2.5263 96 0.7373 -0.0131 0.7373 0.8586
No log 2.5789 98 0.8460 0.0071 0.8460 0.9198
No log 2.6316 100 1.2499 -0.1291 1.2499 1.1180
No log 2.6842 102 1.1845 -0.0638 1.1845 1.0884
No log 2.7368 104 1.1398 -0.1281 1.1398 1.0676
No log 2.7895 106 0.8639 0.0071 0.8639 0.9295
No log 2.8421 108 0.7524 0.0296 0.7524 0.8674
No log 2.8947 110 0.7571 0.0759 0.7571 0.8701
No log 2.9474 112 0.7808 0.0714 0.7808 0.8836
No log 3.0 114 0.8386 -0.0059 0.8386 0.9157
No log 3.0526 116 0.9669 -0.1407 0.9669 0.9833
No log 3.1053 118 1.1570 -0.0837 1.1570 1.0756
No log 3.1579 120 1.2240 -0.0854 1.2240 1.1063
No log 3.2105 122 1.0229 -0.0230 1.0229 1.0114
No log 3.2632 124 1.0093 -0.0870 1.0093 1.0047
No log 3.3158 126 0.9366 -0.1524 0.9366 0.9678
No log 3.3684 128 0.8774 0.0821 0.8774 0.9367
No log 3.4211 130 0.9781 -0.0373 0.9781 0.9890
No log 3.4737 132 0.8862 0.1047 0.8862 0.9414
No log 3.5263 134 0.7799 -0.0578 0.7799 0.8831
No log 3.5789 136 0.7459 0.0541 0.7459 0.8637
No log 3.6316 138 0.7315 -0.0096 0.7315 0.8553
No log 3.6842 140 0.8491 0.0549 0.8491 0.9215
No log 3.7368 142 0.9273 0.0826 0.9273 0.9629
No log 3.7895 144 0.7596 0.1627 0.7596 0.8716
No log 3.8421 146 0.7624 0.0768 0.7624 0.8732
No log 3.8947 148 0.8859 0.0409 0.8859 0.9412
No log 3.9474 150 0.9889 -0.0079 0.9889 0.9944
No log 4.0 152 0.9385 -0.0079 0.9385 0.9688
No log 4.0526 154 0.9042 0.0346 0.9042 0.9509
No log 4.1053 156 0.7919 0.1342 0.7919 0.8899
No log 4.1579 158 0.7940 0.1440 0.7940 0.8911
No log 4.2105 160 0.8748 0.0762 0.8748 0.9353
No log 4.2632 162 0.8893 0.0407 0.8893 0.9430
No log 4.3158 164 1.0095 0.0989 1.0095 1.0048
No log 4.3684 166 0.9339 0.0956 0.9339 0.9664
No log 4.4211 168 1.0411 -0.0052 1.0411 1.0203
No log 4.4737 170 0.9789 0.0490 0.9789 0.9894
No log 4.5263 172 0.9118 -0.1395 0.9118 0.9549
No log 4.5789 174 0.9316 -0.2489 0.9316 0.9652
No log 4.6316 176 0.9741 0.0071 0.9741 0.9870
No log 4.6842 178 1.1701 -0.0532 1.1701 1.0817
No log 4.7368 180 1.0293 0.0711 1.0293 1.0146
No log 4.7895 182 0.8343 -0.1594 0.8343 0.9134
No log 4.8421 184 0.7846 -0.1100 0.7846 0.8858
No log 4.8947 186 0.7638 -0.0725 0.7638 0.8740
No log 4.9474 188 0.9446 0.0748 0.9446 0.9719
No log 5.0 190 1.1222 0.0416 1.1222 1.0594
No log 5.0526 192 0.9055 0.0043 0.9055 0.9516
No log 5.1053 194 0.7509 -0.0215 0.7509 0.8665
No log 5.1579 196 0.7877 -0.0264 0.7877 0.8875
No log 5.2105 198 0.7777 0.0260 0.7777 0.8819
No log 5.2632 200 0.7924 0.0260 0.7924 0.8902
No log 5.3158 202 0.7781 0.0296 0.7781 0.8821
No log 5.3684 204 0.8353 0.0999 0.8353 0.9140
No log 5.4211 206 0.8081 0.1047 0.8081 0.8989
No log 5.4737 208 0.7902 0.0296 0.7902 0.8889
No log 5.5263 210 0.8552 0.0909 0.8552 0.9248
No log 5.5789 212 0.8770 0.0909 0.8770 0.9365
No log 5.6316 214 0.7837 -0.0204 0.7837 0.8853
No log 5.6842 216 0.7782 -0.1100 0.7782 0.8822
No log 5.7368 218 0.7859 -0.1333 0.7859 0.8865
No log 5.7895 220 0.8744 0.1235 0.8744 0.9351
No log 5.8421 222 1.2594 0.0831 1.2594 1.1222
No log 5.8947 224 1.3251 0.0193 1.3251 1.1511
No log 5.9474 226 0.9514 0.0700 0.9514 0.9754
No log 6.0 228 0.8249 0.0185 0.8249 0.9083
No log 6.0526 230 0.7812 -0.1159 0.7812 0.8839
No log 6.1053 232 0.7745 0.1440 0.7745 0.8801
No log 6.1579 234 1.0938 0.0912 1.0938 1.0458
No log 6.2105 236 1.1125 0.0578 1.1125 1.0547
No log 6.2632 238 0.8626 0.1107 0.8626 0.9288
No log 6.3158 240 0.8090 0.0442 0.8090 0.8995
No log 6.3684 242 0.8025 0.0600 0.8025 0.8958
No log 6.4211 244 0.8158 0.0323 0.8158 0.9032
No log 6.4737 246 0.8467 0.0455 0.8467 0.9202
No log 6.5263 248 0.8867 -0.0262 0.8867 0.9416
No log 6.5789 250 0.9865 0.0392 0.9865 0.9932
No log 6.6316 252 1.0479 0.1025 1.0479 1.0237
No log 6.6842 254 0.9773 0.0786 0.9773 0.9886
No log 6.7368 256 0.8930 0.0409 0.8930 0.9450
No log 6.7895 258 0.7975 0.1148 0.7975 0.8930
No log 6.8421 260 0.7502 0.0296 0.7502 0.8661
No log 6.8947 262 0.7638 -0.0113 0.7638 0.8739
No log 6.9474 264 0.8589 0.0109 0.8589 0.9268
No log 7.0 266 0.8815 0.0129 0.8815 0.9389
No log 7.0526 268 0.8121 0.0318 0.8121 0.9012
No log 7.1053 270 0.8220 0.1277 0.8220 0.9066
No log 7.1579 272 0.8583 0.2061 0.8583 0.9265
No log 7.2105 274 0.7782 0.0423 0.7782 0.8822
No log 7.2632 276 0.9878 0.1024 0.9878 0.9939
No log 7.3158 278 1.1857 0.1086 1.1857 1.0889
No log 7.3684 280 0.9896 0.0642 0.9896 0.9948
No log 7.4211 282 0.8673 0.1342 0.8673 0.9313
No log 7.4737 284 0.7961 -0.1278 0.7961 0.8922
No log 7.5263 286 0.8419 0.1352 0.8419 0.9175
No log 7.5789 288 0.8138 -0.0796 0.8138 0.9021
No log 7.6316 290 0.8097 -0.0602 0.8097 0.8999
No log 7.6842 292 0.8593 0.0913 0.8593 0.9270
No log 7.7368 294 0.8303 -0.0616 0.8303 0.9112
No log 7.7895 296 0.8061 -0.1335 0.8061 0.8978
No log 7.8421 298 0.8295 0.0078 0.8295 0.9108
No log 7.8947 300 0.8180 -0.0391 0.8180 0.9044
No log 7.9474 302 0.8376 -0.0630 0.8376 0.9152
No log 8.0 304 0.8919 0.0470 0.8919 0.9444
No log 8.0526 306 0.9028 0.0786 0.9028 0.9502
No log 8.1053 308 0.8341 0.0953 0.8341 0.9133
No log 8.1579 310 0.7747 0.0236 0.7747 0.8802
No log 8.2105 312 0.7809 0.1097 0.7809 0.8837
No log 8.2632 314 0.8296 0.0549 0.8296 0.9108
No log 8.3158 316 0.8470 0.0318 0.8470 0.9203
No log 8.3684 318 0.9355 0.0587 0.9355 0.9672
No log 8.4211 320 0.9530 0.0832 0.9530 0.9762
No log 8.4737 322 0.9888 0.0717 0.9888 0.9944
No log 8.5263 324 1.3672 0.0429 1.3672 1.1693
No log 8.5789 326 1.3459 0.0188 1.3459 1.1601
No log 8.6316 328 1.0320 0.1517 1.0320 1.0159
No log 8.6842 330 0.8023 -0.0549 0.8023 0.8957
No log 8.7368 332 0.7533 -0.0444 0.7533 0.8679
No log 8.7895 334 0.7445 0.0556 0.7445 0.8628
No log 8.8421 336 0.7254 0.1371 0.7254 0.8517
No log 8.8947 338 0.7497 0.0191 0.7497 0.8658
No log 8.9474 340 0.8533 0.0409 0.8533 0.9238
No log 9.0 342 0.8146 0.0476 0.8146 0.9026
No log 9.0526 344 0.7436 0.0191 0.7436 0.8623
No log 9.1053 346 0.7394 0.1379 0.7394 0.8599
No log 9.1579 348 0.7769 0.1362 0.7769 0.8814
No log 9.2105 350 0.8328 0.0600 0.8328 0.9126
No log 9.2632 352 0.9120 0.0913 0.9120 0.9550
No log 9.3158 354 0.8684 0.0551 0.8684 0.9319
No log 9.3684 356 0.8201 -0.0495 0.8201 0.9056
No log 9.4211 358 0.7862 0.0410 0.7862 0.8867
No log 9.4737 360 0.7443 0.0357 0.7443 0.8627
No log 9.5263 362 0.8216 0.1395 0.8216 0.9064
No log 9.5789 364 1.0531 0.0847 1.0531 1.0262
No log 9.6316 366 1.0958 0.1111 1.0958 1.0468
No log 9.6842 368 0.8958 0.0920 0.8958 0.9465
No log 9.7368 370 0.8188 0.1686 0.8188 0.9048
No log 9.7895 372 0.8603 0.1442 0.8603 0.9275
No log 9.8421 374 0.7723 0.1605 0.7723 0.8788
No log 9.8947 376 0.7031 -0.0160 0.7031 0.8385
No log 9.9474 378 0.8062 0.0588 0.8062 0.8979
No log 10.0 380 0.9526 0.0233 0.9526 0.9760
No log 10.0526 382 0.9832 0.0224 0.9832 0.9916
No log 10.1053 384 0.8889 0.0549 0.8889 0.9428
No log 10.1579 386 0.8320 -0.0930 0.8320 0.9121
No log 10.2105 388 0.8621 -0.0741 0.8621 0.9285
No log 10.2632 390 0.8573 -0.1715 0.8573 0.9259
No log 10.3158 392 0.9250 0.0600 0.9250 0.9618
No log 10.3684 394 1.1709 0.0147 1.1709 1.0821
No log 10.4211 396 1.2220 -0.0311 1.2220 1.1054
No log 10.4737 398 1.0495 0.0111 1.0495 1.0245
No log 10.5263 400 0.8327 0.0826 0.8327 0.9125
No log 10.5789 402 0.7268 0.0759 0.7268 0.8525
No log 10.6316 404 0.7128 0.0296 0.7128 0.8443
No log 10.6842 406 0.7442 0.0628 0.7442 0.8626
No log 10.7368 408 0.8294 0.1291 0.8294 0.9107
No log 10.7895 410 0.9652 -0.0236 0.9652 0.9824
No log 10.8421 412 0.9549 -0.0571 0.9549 0.9772
No log 10.8947 414 0.8660 0.1699 0.8660 0.9306
No log 10.9474 416 0.8269 0.0670 0.8269 0.9094
No log 11.0 418 0.8028 0.0670 0.8028 0.8960
No log 11.0526 420 0.8207 0.1727 0.8207 0.9059
No log 11.1053 422 0.9431 0.0157 0.9431 0.9712
No log 11.1579 424 1.1060 0.0651 1.1060 1.0517
No log 11.2105 426 1.1571 0.0878 1.1571 1.0757
No log 11.2632 428 1.0130 0.0378 1.0130 1.0065
No log 11.3158 430 0.8193 0.1660 0.8193 0.9052
No log 11.3684 432 0.7468 0.0759 0.7468 0.8642
No log 11.4211 434 0.7487 0.0296 0.7487 0.8653
No log 11.4737 436 0.7876 0.0999 0.7876 0.8875
No log 11.5263 438 0.9133 0.1243 0.9133 0.9557
No log 11.5789 440 0.9582 0.0748 0.9582 0.9789
No log 11.6316 442 1.0112 -0.0101 1.0112 1.0056
No log 11.6842 444 0.9948 0.0316 0.9948 0.9974
No log 11.7368 446 0.8978 0.0091 0.8978 0.9475
No log 11.7895 448 0.8669 -0.0192 0.8669 0.9310
No log 11.8421 450 0.8786 -0.0271 0.8786 0.9373
No log 11.8947 452 0.8640 0.0608 0.8640 0.9295
No log 11.9474 454 0.8696 0.0525 0.8696 0.9325
No log 12.0 456 0.9537 -0.0079 0.9537 0.9766
No log 12.0526 458 0.9958 -0.0504 0.9958 0.9979
No log 12.1053 460 1.0716 -0.1277 1.0716 1.0352
No log 12.1579 462 1.0676 -0.0936 1.0676 1.0332
No log 12.2105 464 0.9670 -0.0079 0.9670 0.9833
No log 12.2632 466 0.8862 0.0272 0.8862 0.9414
No log 12.3158 468 0.9276 -0.0530 0.9276 0.9631
No log 12.3684 470 0.9087 -0.0193 0.9087 0.9533
No log 12.4211 472 0.8400 -0.0025 0.8400 0.9165
No log 12.4737 474 0.8918 0.0476 0.8918 0.9443
No log 12.5263 476 0.9725 -0.0122 0.9725 0.9862
No log 12.5789 478 0.9582 -0.0393 0.9582 0.9789
No log 12.6316 480 0.9218 0.0245 0.9218 0.9601
No log 12.6842 482 0.9276 -0.0184 0.9276 0.9631
No log 12.7368 484 0.9186 -0.0391 0.9186 0.9584
No log 12.7895 486 0.8895 0.0525 0.8895 0.9431
No log 12.8421 488 0.8972 -0.0459 0.8972 0.9472
No log 12.8947 490 0.9059 -0.0459 0.9059 0.9518
No log 12.9474 492 0.8917 -0.0425 0.8917 0.9443
No log 13.0 494 0.8732 0.0123 0.8732 0.9344
No log 13.0526 496 0.8461 0.1143 0.8461 0.9198
No log 13.1053 498 0.8408 0.0376 0.8408 0.9170
0.357 13.1579 500 0.8354 0.1202 0.8354 0.9140
0.357 13.2105 502 0.8947 -0.0408 0.8947 0.9459
0.357 13.2632 504 0.9812 -0.0033 0.9812 0.9906
0.357 13.3158 506 1.0657 -0.0545 1.0657 1.0324
0.357 13.3684 508 1.0433 -0.0122 1.0433 1.0214
0.357 13.4211 510 1.0235 -0.0183 1.0235 1.0117

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k14_task3_organization

Finetuned
(4023)
this model