ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k6_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6161
  • Qwk: 0.775
  • Mse: 0.6161
  • Rmse: 0.7849

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 6.9321 0.0056 6.9321 2.6329
No log 0.1379 4 4.9687 0.0674 4.9687 2.2290
No log 0.2069 6 3.3971 0.0702 3.3971 1.8431
No log 0.2759 8 3.4812 -0.0109 3.4812 1.8658
No log 0.3448 10 3.5116 -0.0328 3.5116 1.8739
No log 0.4138 12 2.6786 0.0952 2.6786 1.6366
No log 0.4828 14 2.3345 0.0889 2.3345 1.5279
No log 0.5517 16 1.9450 0.2188 1.9450 1.3946
No log 0.6207 18 2.1677 0.25 2.1677 1.4723
No log 0.6897 20 2.0047 0.2797 2.0047 1.4159
No log 0.7586 22 1.8775 0.3262 1.8775 1.3702
No log 0.8276 24 1.7370 0.3429 1.7370 1.3180
No log 0.8966 26 1.8969 0.3784 1.8969 1.3773
No log 0.9655 28 1.7919 0.3784 1.7919 1.3386
No log 1.0345 30 1.8037 0.3624 1.8037 1.3430
No log 1.1034 32 2.4125 0.2295 2.4125 1.5532
No log 1.1724 34 2.7324 0.2073 2.7324 1.6530
No log 1.2414 36 2.5141 0.1538 2.5141 1.5856
No log 1.3103 38 2.0118 0.2483 2.0118 1.4184
No log 1.3793 40 1.7334 0.4412 1.7334 1.3166
No log 1.4483 42 1.6629 0.4412 1.6629 1.2895
No log 1.5172 44 1.4251 0.4889 1.4251 1.1938
No log 1.5862 46 1.3256 0.5103 1.3256 1.1514
No log 1.6552 48 1.8126 0.5056 1.8126 1.3463
No log 1.7241 50 2.7713 0.3445 2.7713 1.6647
No log 1.7931 52 3.1752 0.2984 3.1752 1.7819
No log 1.8621 54 1.9482 0.5048 1.9482 1.3958
No log 1.9310 56 1.0219 0.6225 1.0219 1.0109
No log 2.0 58 1.0082 0.6069 1.0082 1.0041
No log 2.0690 60 1.4511 0.5238 1.4511 1.2046
No log 2.1379 62 1.8859 0.5131 1.8859 1.3733
No log 2.2069 64 2.2367 0.4299 2.2367 1.4956
No log 2.2759 66 1.6558 0.5085 1.6558 1.2868
No log 2.3448 68 1.0807 0.6267 1.0807 1.0396
No log 2.4138 70 0.9524 0.6490 0.9524 0.9759
No log 2.4828 72 0.8033 0.7285 0.8033 0.8963
No log 2.5517 74 0.8011 0.7436 0.8011 0.8950
No log 2.6207 76 0.8355 0.7683 0.8355 0.9141
No log 2.6897 78 0.7834 0.7952 0.7834 0.8851
No log 2.7586 80 0.7943 0.7758 0.7943 0.8912
No log 2.8276 82 0.8392 0.7746 0.8392 0.9161
No log 2.8966 84 0.7760 0.8 0.7760 0.8809
No log 2.9655 86 0.7096 0.8024 0.7096 0.8424
No log 3.0345 88 0.7294 0.7439 0.7294 0.8541
No log 3.1034 90 0.7989 0.6962 0.7989 0.8938
No log 3.1724 92 0.6971 0.8166 0.6971 0.8349
No log 3.2414 94 0.6742 0.8166 0.6742 0.8211
No log 3.3103 96 0.8044 0.7485 0.8044 0.8969
No log 3.3793 98 0.7529 0.7692 0.7529 0.8677
No log 3.4483 100 0.6171 0.8 0.6171 0.7856
No log 3.5172 102 0.6409 0.7952 0.6409 0.8006
No log 3.5862 104 0.7214 0.7674 0.7214 0.8494
No log 3.6552 106 0.9463 0.7079 0.9463 0.9728
No log 3.7241 108 0.8383 0.7363 0.8383 0.9156
No log 3.7931 110 0.6349 0.8045 0.6349 0.7968
No log 3.8621 112 0.6400 0.8022 0.6400 0.8000
No log 3.9310 114 0.7371 0.7619 0.7371 0.8585
No log 4.0 116 0.5931 0.8202 0.5931 0.7701
No log 4.0690 118 0.6046 0.7711 0.6046 0.7776
No log 4.1379 120 0.6385 0.8268 0.6385 0.7991
No log 4.2069 122 0.9736 0.7104 0.9736 0.9867
No log 4.2759 124 0.9674 0.7104 0.9674 0.9836
No log 4.3448 126 0.7073 0.7143 0.7073 0.8410
No log 4.4138 128 0.7371 0.7248 0.7371 0.8585
No log 4.4828 130 0.8304 0.7211 0.8304 0.9113
No log 4.5517 132 1.1460 0.6748 1.1460 1.0705
No log 4.6207 134 1.2495 0.6509 1.2495 1.1178
No log 4.6897 136 0.9381 0.7081 0.9381 0.9686
No log 4.7586 138 0.7097 0.7051 0.7097 0.8425
No log 4.8276 140 0.6578 0.7857 0.6578 0.8111
No log 4.8966 142 0.7224 0.7807 0.7224 0.8499
No log 4.9655 144 0.7781 0.7807 0.7781 0.8821
No log 5.0345 146 0.6294 0.7955 0.6294 0.7933
No log 5.1034 148 0.5735 0.7871 0.5735 0.7573
No log 5.1724 150 0.5941 0.7763 0.5941 0.7708
No log 5.2414 152 0.7224 0.7738 0.7224 0.8499
No log 5.3103 154 1.1106 0.6882 1.1106 1.0538
No log 5.3793 156 1.0077 0.7243 1.0077 1.0038
No log 5.4483 158 0.6772 0.7889 0.6772 0.8229
No log 5.5172 160 0.7106 0.7226 0.7106 0.8430
No log 5.5862 162 0.8871 0.6803 0.8871 0.9419
No log 5.6552 164 0.7496 0.7394 0.7496 0.8658
No log 5.7241 166 0.7254 0.7895 0.7254 0.8517
No log 5.7931 168 0.8119 0.7660 0.8119 0.9010
No log 5.8621 170 0.8263 0.7701 0.8263 0.9090
No log 5.9310 172 0.8374 0.7568 0.8374 0.9151
No log 6.0 174 0.7174 0.7558 0.7174 0.8470
No log 6.0690 176 0.7253 0.7467 0.7253 0.8516
No log 6.1379 178 0.8001 0.6944 0.8001 0.8945
No log 6.2069 180 1.1291 0.6497 1.1291 1.0626
No log 6.2759 182 1.4965 0.4910 1.4965 1.2233
No log 6.3448 184 1.3015 0.5976 1.3015 1.1408
No log 6.4138 186 0.8804 0.6842 0.8804 0.9383
No log 6.4828 188 0.7501 0.7273 0.7501 0.8661
No log 6.5517 190 0.7708 0.7590 0.7708 0.8780
No log 6.6207 192 0.9043 0.7650 0.9043 0.9509
No log 6.6897 194 0.7977 0.7650 0.7977 0.8931
No log 6.7586 196 0.6098 0.7784 0.6098 0.7809
No log 6.8276 198 0.5728 0.7703 0.5728 0.7568
No log 6.8966 200 0.5828 0.7755 0.5828 0.7634
No log 6.9655 202 0.6001 0.7547 0.6001 0.7747
No log 7.0345 204 0.7395 0.7816 0.7395 0.8600
No log 7.1034 206 0.7663 0.7816 0.7663 0.8754
No log 7.1724 208 0.6845 0.7816 0.6845 0.8273
No log 7.2414 210 0.6335 0.8161 0.6335 0.7959
No log 7.3103 212 0.6206 0.8046 0.6206 0.7878
No log 7.3793 214 0.6488 0.7882 0.6488 0.8055
No log 7.4483 216 0.6127 0.7702 0.6127 0.7828
No log 7.5172 218 0.6235 0.7310 0.6235 0.7896
No log 7.5862 220 0.6421 0.7613 0.6421 0.8013
No log 7.6552 222 0.6211 0.7421 0.6211 0.7881
No log 7.7241 224 0.5576 0.7867 0.5576 0.7467
No log 7.7931 226 0.5488 0.7785 0.5488 0.7408
No log 7.8621 228 0.5853 0.775 0.5853 0.7650
No log 7.9310 230 0.7164 0.7861 0.7164 0.8464
No log 8.0 232 0.7438 0.7709 0.7438 0.8624
No log 8.0690 234 0.6042 0.8144 0.6042 0.7773
No log 8.1379 236 0.6063 0.7552 0.6063 0.7787
No log 8.2069 238 0.6509 0.7465 0.6509 0.8068
No log 8.2759 240 0.6504 0.7862 0.6504 0.8065
No log 8.3448 242 0.7441 0.7362 0.7441 0.8626
No log 8.4138 244 0.8075 0.7456 0.8075 0.8986
No log 8.4828 246 0.7254 0.7362 0.7254 0.8517
No log 8.5517 248 0.6713 0.7550 0.6713 0.8193
No log 8.6207 250 0.6580 0.7550 0.6580 0.8112
No log 8.6897 252 0.6764 0.7545 0.6764 0.8224
No log 8.7586 254 0.7547 0.7778 0.7547 0.8687
No log 8.8276 256 0.6762 0.7826 0.6762 0.8223
No log 8.8966 258 0.5661 0.8156 0.5661 0.7524
No log 8.9655 260 0.5605 0.7778 0.5605 0.7487
No log 9.0345 262 0.5619 0.7673 0.5619 0.7496
No log 9.1034 264 0.5956 0.8256 0.5956 0.7717
No log 9.1724 266 0.6685 0.7977 0.6685 0.8176
No log 9.2414 268 0.7365 0.7453 0.7365 0.8582
No log 9.3103 270 0.6881 0.7383 0.6881 0.8295
No log 9.3793 272 0.6678 0.7310 0.6678 0.8172
No log 9.4483 274 0.6848 0.7042 0.6848 0.8275
No log 9.5172 276 0.6856 0.7403 0.6856 0.8280
No log 9.5862 278 0.6750 0.75 0.6750 0.8216
No log 9.6552 280 0.6677 0.7904 0.6677 0.8171
No log 9.7241 282 0.6772 0.7952 0.6772 0.8229
No log 9.7931 284 0.7185 0.7683 0.7185 0.8476
No log 9.8621 286 0.8281 0.7630 0.8281 0.9100
No log 9.9310 288 0.9360 0.7314 0.9360 0.9674
No log 10.0 290 0.8928 0.7442 0.8928 0.9449
No log 10.0690 292 0.7694 0.7389 0.7694 0.8771
No log 10.1379 294 0.7039 0.7285 0.7039 0.8390
No log 10.2069 296 0.6871 0.7172 0.6871 0.8289
No log 10.2759 298 0.6563 0.7211 0.6563 0.8101
No log 10.3448 300 0.6725 0.7403 0.6725 0.8201
No log 10.4138 302 0.8164 0.7578 0.8164 0.9035
No log 10.4828 304 0.9181 0.7598 0.9181 0.9582
No log 10.5517 306 0.8085 0.7738 0.8085 0.8992
No log 10.6207 308 0.6688 0.7308 0.6688 0.8178
No log 10.6897 310 0.6049 0.7397 0.6049 0.7777
No log 10.7586 312 0.6330 0.7324 0.6330 0.7956
No log 10.8276 314 0.6424 0.7324 0.6424 0.8015
No log 10.8966 316 0.6369 0.7417 0.6369 0.7981
No log 10.9655 318 0.7587 0.7412 0.7587 0.8711
No log 11.0345 320 0.8404 0.7135 0.8404 0.9167
No log 11.1034 322 0.8294 0.7262 0.8294 0.9107
No log 11.1724 324 0.7060 0.7389 0.7060 0.8403
No log 11.2414 326 0.6560 0.7448 0.6560 0.8099
No log 11.3103 328 0.6595 0.75 0.6595 0.8121
No log 11.3793 330 0.6800 0.72 0.6800 0.8246
No log 11.4483 332 0.7674 0.7456 0.7674 0.8760
No log 11.5172 334 0.8589 0.7333 0.8589 0.9268
No log 11.5862 336 0.7496 0.7582 0.7496 0.8658
No log 11.6552 338 0.5926 0.8114 0.5926 0.7698
No log 11.7241 340 0.5606 0.7976 0.5606 0.7487
No log 11.7931 342 0.5682 0.8187 0.5682 0.7538
No log 11.8621 344 0.6120 0.8023 0.6120 0.7823
No log 11.9310 346 0.6286 0.7935 0.6286 0.7929
No log 12.0 348 0.5737 0.8295 0.5737 0.7574
No log 12.0690 350 0.5485 0.8263 0.5485 0.7406
No log 12.1379 352 0.5562 0.7949 0.5562 0.7458
No log 12.2069 354 0.6105 0.7973 0.6105 0.7814
No log 12.2759 356 0.5700 0.7632 0.5700 0.7550
No log 12.3448 358 0.5415 0.8304 0.5415 0.7358
No log 12.4138 360 0.6276 0.7556 0.6276 0.7922
No log 12.4828 362 0.6739 0.7556 0.6739 0.8209
No log 12.5517 364 0.6709 0.7556 0.6709 0.8191
No log 12.6207 366 0.6378 0.7727 0.6378 0.7986
No log 12.6897 368 0.5847 0.8272 0.5847 0.7647
No log 12.7586 370 0.5942 0.7895 0.5942 0.7709
No log 12.8276 372 0.6224 0.7682 0.6224 0.7889
No log 12.8966 374 0.6314 0.7947 0.6314 0.7946
No log 12.9655 376 0.6624 0.7792 0.6624 0.8139
No log 13.0345 378 0.6793 0.7792 0.6793 0.8242
No log 13.1034 380 0.6622 0.7848 0.6622 0.8138
No log 13.1724 382 0.6306 0.7950 0.6306 0.7941
No log 13.2414 384 0.6230 0.7673 0.6230 0.7893
No log 13.3103 386 0.6260 0.7758 0.6260 0.7912
No log 13.3793 388 0.6429 0.7738 0.6429 0.8018
No log 13.4483 390 0.6614 0.7665 0.6614 0.8133
No log 13.5172 392 0.7515 0.7442 0.7515 0.8669
No log 13.5862 394 0.7452 0.7683 0.7452 0.8632
No log 13.6552 396 0.6639 0.7625 0.6639 0.8148
No log 13.7241 398 0.6328 0.7333 0.6328 0.7955
No log 13.7931 400 0.6560 0.7273 0.6560 0.8100
No log 13.8621 402 0.6537 0.7361 0.6537 0.8085
No log 13.9310 404 0.6648 0.7975 0.6648 0.8153
No log 14.0 406 0.7698 0.7692 0.7698 0.8774
No log 14.0690 408 0.7961 0.7771 0.7961 0.8922
No log 14.1379 410 0.7244 0.7771 0.7244 0.8511
No log 14.2069 412 0.6445 0.8047 0.6445 0.8028
No log 14.2759 414 0.6073 0.7950 0.6073 0.7793
No log 14.3448 416 0.6045 0.7432 0.6045 0.7775
No log 14.4138 418 0.6383 0.7162 0.6383 0.7990
No log 14.4828 420 0.6922 0.7285 0.6922 0.8320
No log 14.5517 422 0.7052 0.7075 0.7052 0.8398
No log 14.6207 424 0.6521 0.7448 0.6521 0.8075
No log 14.6897 426 0.5725 0.8054 0.5725 0.7566
No log 14.7586 428 0.5868 0.75 0.5868 0.7660
No log 14.8276 430 0.5968 0.7552 0.5968 0.7725
No log 14.8966 432 0.5461 0.7552 0.5461 0.7390
No log 14.9655 434 0.5206 0.7799 0.5206 0.7215
No log 15.0345 436 0.6874 0.7727 0.6874 0.8291
No log 15.1034 438 0.8257 0.7374 0.8257 0.9087
No log 15.1724 440 0.7031 0.7657 0.7031 0.8385
No log 15.2414 442 0.6212 0.7882 0.6212 0.7882
No log 15.3103 444 0.6430 0.7882 0.6430 0.8019
No log 15.3793 446 0.6462 0.7799 0.6462 0.8039
No log 15.4483 448 0.7168 0.7451 0.7168 0.8466
No log 15.5172 450 0.7248 0.7403 0.7248 0.8514
No log 15.5862 452 0.7217 0.7636 0.7217 0.8495
No log 15.6552 454 0.7118 0.7657 0.7118 0.8437
No log 15.7241 456 0.6322 0.8023 0.6322 0.7951
No log 15.7931 458 0.5808 0.8439 0.5808 0.7621
No log 15.8621 460 0.5785 0.8187 0.5785 0.7606
No log 15.9310 462 0.5888 0.8439 0.5888 0.7674
No log 16.0 464 0.6110 0.8 0.6110 0.7817
No log 16.0690 466 0.5776 0.8140 0.5776 0.7600
No log 16.1379 468 0.5590 0.8263 0.5590 0.7476
No log 16.2069 470 0.5632 0.8171 0.5632 0.7504
No log 16.2759 472 0.5926 0.8118 0.5926 0.7698
No log 16.3448 474 0.5932 0.8118 0.5932 0.7702
No log 16.4138 476 0.5836 0.8118 0.5836 0.7639
No log 16.4828 478 0.6009 0.7952 0.6009 0.7752
No log 16.5517 480 0.6258 0.7516 0.6258 0.7911
No log 16.6207 482 0.6658 0.7237 0.6658 0.8160
No log 16.6897 484 0.6480 0.7285 0.6480 0.8050
No log 16.7586 486 0.6323 0.7826 0.6323 0.7952
No log 16.8276 488 0.6191 0.7826 0.6191 0.7868
No log 16.8966 490 0.6242 0.8193 0.6242 0.7901
No log 16.9655 492 0.6047 0.8 0.6047 0.7776
No log 17.0345 494 0.6153 0.8193 0.6153 0.7844
No log 17.1034 496 0.6175 0.8193 0.6175 0.7858
No log 17.1724 498 0.6387 0.8121 0.6387 0.7992
0.3506 17.2414 500 0.6494 0.8049 0.6494 0.8059
0.3506 17.3103 502 0.6839 0.7758 0.6839 0.8270
0.3506 17.3793 504 0.7515 0.7485 0.7515 0.8669
0.3506 17.4483 506 0.7166 0.7816 0.7166 0.8466
0.3506 17.5172 508 0.6125 0.8140 0.6125 0.7826
0.3506 17.5862 510 0.5433 0.8121 0.5433 0.7371
0.3506 17.6552 512 0.5254 0.7692 0.5254 0.7248
0.3506 17.7241 514 0.5474 0.7586 0.5474 0.7398
0.3506 17.7931 516 0.5510 0.7586 0.5510 0.7423
0.3506 17.8621 518 0.5529 0.7871 0.5529 0.7436
0.3506 17.9310 520 0.6161 0.775 0.6161 0.7849

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k6_task1_organization

Finetuned
(4023)
this model