ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k8_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6476
  • Qwk: 0.6021
  • Mse: 0.6476
  • Rmse: 0.8047

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 4.4109 -0.0061 4.4109 2.1002
No log 0.1429 4 2.5660 -0.0111 2.5660 1.6019
No log 0.2143 6 1.6001 -0.0180 1.6001 1.2649
No log 0.2857 8 1.1051 0.2687 1.1051 1.0512
No log 0.3571 10 1.1263 0.1589 1.1263 1.0613
No log 0.4286 12 1.2105 0.0369 1.2105 1.1002
No log 0.5 14 1.5488 0.0250 1.5488 1.2445
No log 0.5714 16 1.9483 0.0850 1.9483 1.3958
No log 0.6429 18 1.6629 0.0399 1.6629 1.2896
No log 0.7143 20 1.2901 0.0232 1.2901 1.1358
No log 0.7857 22 1.1964 0.0232 1.1964 1.0938
No log 0.8571 24 1.0985 0.1564 1.0985 1.0481
No log 0.9286 26 1.0465 0.2615 1.0465 1.0230
No log 1.0 28 1.0551 0.1361 1.0551 1.0272
No log 1.0714 30 1.0756 0.1805 1.0756 1.0371
No log 1.1429 32 1.1549 0.0996 1.1549 1.0747
No log 1.2143 34 1.2069 0.0760 1.2069 1.0986
No log 1.2857 36 1.2906 0.0760 1.2906 1.1360
No log 1.3571 38 1.3265 0.0878 1.3265 1.1517
No log 1.4286 40 1.1166 0.1764 1.1166 1.0567
No log 1.5 42 1.0521 0.2004 1.0521 1.0257
No log 1.5714 44 1.1729 0.2479 1.1729 1.0830
No log 1.6429 46 1.5689 0.1893 1.5689 1.2526
No log 1.7143 48 1.5978 0.1893 1.5978 1.2640
No log 1.7857 50 1.2773 0.2632 1.2773 1.1302
No log 1.8571 52 1.1002 0.2371 1.1002 1.0489
No log 1.9286 54 0.8433 0.4908 0.8433 0.9183
No log 2.0 56 0.8311 0.4345 0.8311 0.9116
No log 2.0714 58 0.8237 0.4524 0.8237 0.9076
No log 2.1429 60 0.9035 0.3902 0.9035 0.9505
No log 2.2143 62 0.8848 0.4353 0.8848 0.9406
No log 2.2857 64 0.8690 0.4755 0.8690 0.9322
No log 2.3571 66 1.1941 0.3323 1.1941 1.0927
No log 2.4286 68 1.2536 0.2727 1.2536 1.1196
No log 2.5 70 0.9957 0.4606 0.9957 0.9979
No log 2.5714 72 0.8846 0.4368 0.8846 0.9405
No log 2.6429 74 1.1122 0.3584 1.1122 1.0546
No log 2.7143 76 1.0899 0.3212 1.0899 1.0440
No log 2.7857 78 0.9153 0.4254 0.9153 0.9567
No log 2.8571 80 0.9580 0.4379 0.9580 0.9788
No log 2.9286 82 1.0635 0.3654 1.0635 1.0312
No log 3.0 84 1.0267 0.4218 1.0267 1.0133
No log 3.0714 86 0.8962 0.4388 0.8962 0.9467
No log 3.1429 88 0.8455 0.4352 0.8455 0.9195
No log 3.2143 90 1.0958 0.2843 1.0958 1.0468
No log 3.2857 92 1.1938 0.3195 1.1938 1.0926
No log 3.3571 94 0.9421 0.4070 0.9421 0.9706
No log 3.4286 96 0.8219 0.4865 0.8219 0.9066
No log 3.5 98 0.9992 0.3986 0.9992 0.9996
No log 3.5714 100 0.9990 0.3986 0.9990 0.9995
No log 3.6429 102 0.8106 0.4616 0.8106 0.9004
No log 3.7143 104 0.8153 0.5312 0.8153 0.9029
No log 3.7857 106 0.9565 0.4561 0.9565 0.9780
No log 3.8571 108 0.8677 0.5309 0.8677 0.9315
No log 3.9286 110 0.8197 0.5395 0.8197 0.9054
No log 4.0 112 0.7530 0.5585 0.7530 0.8677
No log 4.0714 114 0.7496 0.5919 0.7496 0.8658
No log 4.1429 116 0.7815 0.5737 0.7815 0.8840
No log 4.2143 118 0.8798 0.4937 0.8798 0.9380
No log 4.2857 120 1.0558 0.4155 1.0558 1.0275
No log 4.3571 122 0.9547 0.3935 0.9547 0.9771
No log 4.4286 124 0.7225 0.5561 0.7225 0.8500
No log 4.5 126 0.7834 0.5788 0.7834 0.8851
No log 4.5714 128 1.0698 0.4763 1.0698 1.0343
No log 4.6429 130 1.0625 0.4763 1.0625 1.0308
No log 4.7143 132 0.8195 0.56 0.8195 0.9053
No log 4.7857 134 0.6402 0.6144 0.6402 0.8002
No log 4.8571 136 0.8302 0.5054 0.8302 0.9111
No log 4.9286 138 0.8868 0.5363 0.8868 0.9417
No log 5.0 140 0.7229 0.5932 0.7229 0.8503
No log 5.0714 142 0.6270 0.6398 0.6270 0.7918
No log 5.1429 144 0.6697 0.5787 0.6697 0.8184
No log 5.2143 146 0.6510 0.5710 0.6510 0.8069
No log 5.2857 148 0.6399 0.6039 0.6399 0.7999
No log 5.3571 150 0.7288 0.6082 0.7288 0.8537
No log 5.4286 152 0.7109 0.6166 0.7109 0.8432
No log 5.5 154 0.6616 0.6350 0.6616 0.8134
No log 5.5714 156 0.8440 0.5978 0.8440 0.9187
No log 5.6429 158 0.9825 0.5135 0.9825 0.9912
No log 5.7143 160 0.8668 0.5476 0.8668 0.9310
No log 5.7857 162 0.7010 0.5657 0.7010 0.8373
No log 5.8571 164 0.7203 0.5714 0.7203 0.8487
No log 5.9286 166 0.7641 0.5093 0.7641 0.8741
No log 6.0 168 0.8105 0.4990 0.8105 0.9003
No log 6.0714 170 0.8031 0.4864 0.8031 0.8962
No log 6.1429 172 0.7446 0.5659 0.7446 0.8629
No log 6.2143 174 0.7196 0.4862 0.7196 0.8483
No log 6.2857 176 0.7584 0.5070 0.7584 0.8708
No log 6.3571 178 0.7567 0.4946 0.7567 0.8699
No log 6.4286 180 0.6969 0.4478 0.6969 0.8348
No log 6.5 182 0.6661 0.5868 0.6661 0.8161
No log 6.5714 184 0.6799 0.5854 0.6799 0.8246
No log 6.6429 186 0.6589 0.5868 0.6589 0.8117
No log 6.7143 188 0.6603 0.5248 0.6603 0.8126
No log 6.7857 190 0.6704 0.5657 0.6704 0.8188
No log 6.8571 192 0.6813 0.5731 0.6813 0.8254
No log 6.9286 194 0.7323 0.6215 0.7323 0.8557
No log 7.0 196 0.8246 0.5402 0.8246 0.9081
No log 7.0714 198 0.8344 0.5577 0.8344 0.9134
No log 7.1429 200 0.7302 0.5837 0.7302 0.8545
No log 7.2143 202 0.6923 0.6157 0.6923 0.8320
No log 7.2857 204 0.6964 0.5522 0.6964 0.8345
No log 7.3571 206 0.6865 0.5996 0.6865 0.8285
No log 7.4286 208 0.6965 0.5171 0.6965 0.8346
No log 7.5 210 0.7664 0.4998 0.7664 0.8754
No log 7.5714 212 0.7731 0.4450 0.7731 0.8793
No log 7.6429 214 0.7412 0.4675 0.7412 0.8609
No log 7.7143 216 0.7277 0.4893 0.7277 0.8531
No log 7.7857 218 0.7033 0.5002 0.7033 0.8386
No log 7.8571 220 0.7124 0.6060 0.7124 0.8440
No log 7.9286 222 0.7667 0.5895 0.7667 0.8756
No log 8.0 224 0.7526 0.6017 0.7526 0.8675
No log 8.0714 226 0.6907 0.6224 0.6907 0.8311
No log 8.1429 228 0.6829 0.5797 0.6829 0.8264
No log 8.2143 230 0.6726 0.6120 0.6726 0.8201
No log 8.2857 232 0.7198 0.6229 0.7198 0.8484
No log 8.3571 234 0.7679 0.6065 0.7679 0.8763
No log 8.4286 236 0.7065 0.5940 0.7065 0.8406
No log 8.5 238 0.6545 0.5923 0.6545 0.8090
No log 8.5714 240 0.6481 0.6028 0.6481 0.8050
No log 8.6429 242 0.6662 0.5542 0.6662 0.8162
No log 8.7143 244 0.7752 0.5571 0.7752 0.8805
No log 8.7857 246 0.8921 0.5154 0.8921 0.9445
No log 8.8571 248 0.7604 0.5220 0.7604 0.8720
No log 8.9286 250 0.6556 0.6442 0.6556 0.8097
No log 9.0 252 0.7310 0.5660 0.7310 0.8550
No log 9.0714 254 0.7657 0.5870 0.7657 0.8751
No log 9.1429 256 0.7065 0.6160 0.7065 0.8405
No log 9.2143 258 0.6454 0.6259 0.6454 0.8034
No log 9.2857 260 0.6862 0.5217 0.6862 0.8284
No log 9.3571 262 0.6910 0.5327 0.6910 0.8313
No log 9.4286 264 0.6693 0.5542 0.6693 0.8181
No log 9.5 266 0.6559 0.5692 0.6559 0.8099
No log 9.5714 268 0.7294 0.6029 0.7294 0.8541
No log 9.6429 270 0.7541 0.6029 0.7541 0.8684
No log 9.7143 272 0.7409 0.5366 0.7409 0.8608
No log 9.7857 274 0.8498 0.4767 0.8498 0.9219
No log 9.8571 276 0.9584 0.3702 0.9584 0.9790
No log 9.9286 278 0.9372 0.4412 0.9372 0.9681
No log 10.0 280 0.8397 0.4513 0.8397 0.9163
No log 10.0714 282 0.8177 0.4502 0.8177 0.9043
No log 10.1429 284 0.8605 0.4750 0.8605 0.9276
No log 10.2143 286 0.9211 0.4656 0.9211 0.9598
No log 10.2857 288 0.8834 0.4641 0.8834 0.9399
No log 10.3571 290 0.8304 0.4971 0.8304 0.9113
No log 10.4286 292 0.7856 0.4971 0.7856 0.8864
No log 10.5 294 0.7607 0.5060 0.7607 0.8722
No log 10.5714 296 0.7840 0.5513 0.7840 0.8854
No log 10.6429 298 0.7628 0.5717 0.7628 0.8734
No log 10.7143 300 0.7410 0.6025 0.7410 0.8608
No log 10.7857 302 0.6772 0.5923 0.6772 0.8229
No log 10.8571 304 0.6779 0.6207 0.6779 0.8233
No log 10.9286 306 0.6755 0.6102 0.6755 0.8219
No log 11.0 308 0.6505 0.5471 0.6505 0.8065
No log 11.0714 310 0.7276 0.5945 0.7276 0.8530
No log 11.1429 312 0.7863 0.6194 0.7863 0.8868
No log 11.2143 314 0.7339 0.6195 0.7339 0.8567
No log 11.2857 316 0.6767 0.5599 0.6767 0.8226
No log 11.3571 318 0.6873 0.5443 0.6873 0.8290
No log 11.4286 320 0.6962 0.6083 0.6962 0.8344
No log 11.5 322 0.6730 0.5808 0.6730 0.8204
No log 11.5714 324 0.6657 0.5327 0.6657 0.8159
No log 11.6429 326 0.7075 0.6047 0.7075 0.8412
No log 11.7143 328 0.8849 0.5745 0.8849 0.9407
No log 11.7857 330 0.8955 0.5489 0.8955 0.9463
No log 11.8571 332 0.7312 0.6175 0.7312 0.8551
No log 11.9286 334 0.6478 0.5671 0.6478 0.8049
No log 12.0 336 0.6534 0.5224 0.6534 0.8083
No log 12.0714 338 0.6646 0.5329 0.6646 0.8153
No log 12.1429 340 0.6761 0.5819 0.6761 0.8222
No log 12.2143 342 0.7194 0.5854 0.7194 0.8482
No log 12.2857 344 0.6968 0.5854 0.6968 0.8348
No log 12.3571 346 0.6694 0.5736 0.6694 0.8182
No log 12.4286 348 0.6501 0.5618 0.6501 0.8063
No log 12.5 350 0.6638 0.5224 0.6638 0.8147
No log 12.5714 352 0.6674 0.5224 0.6674 0.8170
No log 12.6429 354 0.6589 0.5139 0.6589 0.8117
No log 12.7143 356 0.6702 0.5084 0.6702 0.8187
No log 12.7857 358 0.7154 0.5593 0.7154 0.8458
No log 12.8571 360 0.7258 0.5718 0.7258 0.8519
No log 12.9286 362 0.6957 0.5342 0.6957 0.8341
No log 13.0 364 0.6564 0.5287 0.6564 0.8102
No log 13.0714 366 0.6536 0.5399 0.6536 0.8085
No log 13.1429 368 0.6688 0.5425 0.6688 0.8178
No log 13.2143 370 0.6867 0.5662 0.6867 0.8287
No log 13.2857 372 0.7467 0.5364 0.7467 0.8641
No log 13.3571 374 0.7442 0.5476 0.7442 0.8627
No log 13.4286 376 0.6920 0.5659 0.6920 0.8319
No log 13.5 378 0.6823 0.4918 0.6823 0.8260
No log 13.5714 380 0.7045 0.4947 0.7045 0.8394
No log 13.6429 382 0.7289 0.5228 0.7289 0.8537
No log 13.7143 384 0.7771 0.5256 0.7771 0.8815
No log 13.7857 386 0.7662 0.5243 0.7662 0.8753
No log 13.8571 388 0.7012 0.5060 0.7012 0.8374
No log 13.9286 390 0.6703 0.4883 0.6703 0.8187
No log 14.0 392 0.6556 0.5434 0.6556 0.8097
No log 14.0714 394 0.6403 0.5443 0.6403 0.8002
No log 14.1429 396 0.6457 0.5843 0.6457 0.8035
No log 14.2143 398 0.6622 0.6184 0.6622 0.8137
No log 14.2857 400 0.6384 0.5237 0.6384 0.7990
No log 14.3571 402 0.6473 0.5637 0.6473 0.8046
No log 14.4286 404 0.6695 0.5626 0.6695 0.8182
No log 14.5 406 0.6615 0.5425 0.6615 0.8133
No log 14.5714 408 0.6249 0.5577 0.6249 0.7905
No log 14.6429 410 0.6194 0.5701 0.6194 0.7870
No log 14.7143 412 0.6144 0.5913 0.6144 0.7839
No log 14.7857 414 0.6116 0.5913 0.6116 0.7820
No log 14.8571 416 0.6078 0.5990 0.6078 0.7796
No log 14.9286 418 0.6088 0.5929 0.6088 0.7803
No log 15.0 420 0.6105 0.6581 0.6105 0.7814
No log 15.0714 422 0.6012 0.5929 0.6012 0.7754
No log 15.1429 424 0.6094 0.5929 0.6094 0.7806
No log 15.2143 426 0.6220 0.5880 0.6220 0.7887
No log 15.2857 428 0.6318 0.5467 0.6318 0.7948
No log 15.3571 430 0.6492 0.4873 0.6492 0.8057
No log 15.4286 432 0.6623 0.4626 0.6623 0.8138
No log 15.5 434 0.6852 0.4416 0.6852 0.8278
No log 15.5714 436 0.6721 0.4520 0.6721 0.8198
No log 15.6429 438 0.6427 0.5232 0.6427 0.8017
No log 15.7143 440 0.6437 0.5676 0.6437 0.8023
No log 15.7857 442 0.6255 0.5909 0.6255 0.7909
No log 15.8571 444 0.6088 0.6007 0.6088 0.7802
No log 15.9286 446 0.6023 0.6374 0.6023 0.7761
No log 16.0 448 0.6078 0.6713 0.6078 0.7796
No log 16.0714 450 0.6217 0.6677 0.6217 0.7885
No log 16.1429 452 0.6153 0.6554 0.6153 0.7844
No log 16.2143 454 0.5925 0.6880 0.5925 0.7698
No log 16.2857 456 0.5935 0.6374 0.5935 0.7704
No log 16.3571 458 0.6257 0.5353 0.6257 0.7910
No log 16.4286 460 0.6673 0.5245 0.6673 0.8169
No log 16.5 462 0.6671 0.5459 0.6671 0.8168
No log 16.5714 464 0.6171 0.5966 0.6171 0.7855
No log 16.6429 466 0.5923 0.5725 0.5923 0.7696
No log 16.7143 468 0.5984 0.6018 0.5984 0.7736
No log 16.7857 470 0.6097 0.6432 0.6097 0.7809
No log 16.8571 472 0.6089 0.5701 0.6089 0.7803
No log 16.9286 474 0.6267 0.5192 0.6267 0.7916
No log 17.0 476 0.7122 0.5928 0.7122 0.8439
No log 17.0714 478 0.7333 0.6259 0.7333 0.8563
No log 17.1429 480 0.6935 0.6140 0.6935 0.8328
No log 17.2143 482 0.6097 0.5542 0.6097 0.7809
No log 17.2857 484 0.6005 0.6035 0.6005 0.7749
No log 17.3571 486 0.6044 0.5472 0.6044 0.7774
No log 17.4286 488 0.5955 0.6084 0.5955 0.7717
No log 17.5 490 0.5931 0.6067 0.5931 0.7701
No log 17.5714 492 0.5957 0.6067 0.5957 0.7718
No log 17.6429 494 0.5958 0.6067 0.5958 0.7719
No log 17.7143 496 0.5970 0.6057 0.5970 0.7727
No log 17.7857 498 0.5979 0.6249 0.5979 0.7733
0.2647 17.8571 500 0.6017 0.6046 0.6017 0.7757
0.2647 17.9286 502 0.6051 0.6046 0.6051 0.7779
0.2647 18.0 504 0.6134 0.5359 0.6134 0.7832
0.2647 18.0714 506 0.6424 0.5898 0.6424 0.8015
0.2647 18.1429 508 0.6576 0.5898 0.6576 0.8109
0.2647 18.2143 510 0.6476 0.6021 0.6476 0.8047

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k8_task5_organization

Finetuned
(4019)
this model