ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k20_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6901
  • Qwk: 0.4727
  • Mse: 0.6901
  • Rmse: 0.8307

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.02 2 4.2735 0.0035 4.2735 2.0673
No log 0.04 4 2.5890 0.0155 2.5890 1.6090
No log 0.06 6 1.4106 0.0760 1.4106 1.1877
No log 0.08 8 1.1023 0.1631 1.1023 1.0499
No log 0.1 10 1.3042 0.0369 1.3042 1.1420
No log 0.12 12 2.2425 0.0034 2.2425 1.4975
No log 0.14 14 2.5404 0.0192 2.5404 1.5939
No log 0.16 16 1.9704 0.0916 1.9704 1.4037
No log 0.18 18 1.2635 0.1255 1.2635 1.1241
No log 0.2 20 0.9459 0.3221 0.9459 0.9726
No log 0.22 22 0.9524 0.3106 0.9524 0.9759
No log 0.24 24 0.9411 0.3410 0.9411 0.9701
No log 0.26 26 0.9402 0.3499 0.9402 0.9697
No log 0.28 28 0.9386 0.3026 0.9386 0.9688
No log 0.3 30 0.9225 0.2903 0.9225 0.9605
No log 0.32 32 0.8401 0.3590 0.8401 0.9165
No log 0.34 34 0.8504 0.3772 0.8504 0.9222
No log 0.36 36 0.8255 0.4157 0.8255 0.9086
No log 0.38 38 0.6999 0.5694 0.6999 0.8366
No log 0.4 40 0.6595 0.6584 0.6595 0.8121
No log 0.42 42 0.6611 0.6584 0.6611 0.8131
No log 0.44 44 0.6964 0.6420 0.6964 0.8345
No log 0.46 46 0.6713 0.6109 0.6713 0.8193
No log 0.48 48 0.6637 0.6358 0.6637 0.8147
No log 0.5 50 1.1189 0.4236 1.1189 1.0578
No log 0.52 52 1.2182 0.4284 1.2182 1.1037
No log 0.54 54 0.8668 0.5739 0.8668 0.9310
No log 0.56 56 0.6296 0.6224 0.6296 0.7935
No log 0.58 58 0.6302 0.6414 0.6302 0.7939
No log 0.6 60 0.7335 0.6266 0.7335 0.8564
No log 0.62 62 1.3532 0.3650 1.3532 1.1633
No log 0.64 64 1.6548 0.2926 1.6548 1.2864
No log 0.66 66 1.2959 0.3754 1.2959 1.1384
No log 0.68 68 0.7052 0.6157 0.7052 0.8398
No log 0.7 70 0.7102 0.6230 0.7102 0.8427
No log 0.72 72 0.7934 0.6140 0.7934 0.8908
No log 0.74 74 0.6380 0.6981 0.6380 0.7988
No log 0.76 76 0.6300 0.6148 0.6300 0.7937
No log 0.78 78 0.6351 0.5726 0.6351 0.7969
No log 0.8 80 0.6792 0.5770 0.6792 0.8241
No log 0.82 82 0.7062 0.5012 0.7062 0.8404
No log 0.84 84 0.6656 0.5442 0.6656 0.8158
No log 0.86 86 0.6125 0.5638 0.6125 0.7826
No log 0.88 88 0.6085 0.6259 0.6085 0.7800
No log 0.9 90 0.6123 0.6288 0.6123 0.7825
No log 0.92 92 0.6193 0.6614 0.6193 0.7869
No log 0.94 94 0.6213 0.6345 0.6213 0.7882
No log 0.96 96 0.6173 0.6382 0.6173 0.7857
No log 0.98 98 0.6097 0.6266 0.6097 0.7808
No log 1.0 100 0.6508 0.5748 0.6508 0.8067
No log 1.02 102 0.7425 0.5845 0.7425 0.8617
No log 1.04 104 0.8138 0.5730 0.8138 0.9021
No log 1.06 106 0.7044 0.6144 0.7044 0.8393
No log 1.08 108 0.5978 0.6572 0.5978 0.7732
No log 1.1 110 0.6740 0.6179 0.6740 0.8210
No log 1.12 112 0.6568 0.6412 0.6568 0.8105
No log 1.1400 114 0.6153 0.7049 0.6153 0.7844
No log 1.16 116 0.6186 0.6622 0.6186 0.7865
No log 1.18 118 0.6306 0.6773 0.6306 0.7941
No log 1.2 120 0.6211 0.6594 0.6211 0.7881
No log 1.22 122 0.6282 0.6433 0.6282 0.7926
No log 1.24 124 0.6636 0.6299 0.6636 0.8146
No log 1.26 126 0.6522 0.6223 0.6522 0.8076
No log 1.28 128 0.5996 0.6417 0.5996 0.7743
No log 1.3 130 0.6443 0.6617 0.6443 0.8027
No log 1.32 132 0.6642 0.6617 0.6642 0.8150
No log 1.34 134 0.6023 0.6805 0.6023 0.7761
No log 1.3600 136 0.5982 0.6408 0.5982 0.7735
No log 1.38 138 0.6020 0.6694 0.6020 0.7759
No log 1.4 140 0.5926 0.6714 0.5926 0.7698
No log 1.42 142 0.6131 0.6797 0.6131 0.7830
No log 1.44 144 0.6084 0.6693 0.6084 0.7800
No log 1.46 146 0.5962 0.6693 0.5962 0.7722
No log 1.48 148 0.5773 0.6641 0.5773 0.7598
No log 1.5 150 0.5923 0.6364 0.5923 0.7696
No log 1.52 152 0.5886 0.6690 0.5886 0.7672
No log 1.54 154 0.5854 0.6316 0.5854 0.7651
No log 1.56 156 0.6399 0.6283 0.6399 0.7999
No log 1.58 158 0.5978 0.6718 0.5978 0.7731
No log 1.6 160 0.5685 0.6617 0.5685 0.7540
No log 1.62 162 0.7217 0.6624 0.7217 0.8495
No log 1.6400 164 0.6842 0.6597 0.6842 0.8271
No log 1.6600 166 0.6040 0.6388 0.6040 0.7772
No log 1.6800 168 0.5798 0.6284 0.5798 0.7615
No log 1.7 170 0.5975 0.6078 0.5975 0.7730
No log 1.72 172 0.6095 0.5882 0.6095 0.7807
No log 1.74 174 0.6110 0.6307 0.6110 0.7817
No log 1.76 176 0.6168 0.6814 0.6168 0.7854
No log 1.78 178 0.6472 0.6724 0.6472 0.8045
No log 1.8 180 0.6655 0.6485 0.6655 0.8158
No log 1.8200 182 0.6557 0.6623 0.6557 0.8098
No log 1.8400 184 0.6318 0.6673 0.6318 0.7949
No log 1.8600 186 0.6262 0.6272 0.6262 0.7913
No log 1.88 188 0.6155 0.6461 0.6155 0.7845
No log 1.9 190 0.6435 0.6019 0.6435 0.8022
No log 1.92 192 0.6091 0.6812 0.6091 0.7805
No log 1.94 194 0.6140 0.6783 0.6140 0.7836
No log 1.96 196 0.7805 0.6128 0.7805 0.8835
No log 1.98 198 0.7662 0.6128 0.7662 0.8753
No log 2.0 200 0.6191 0.6365 0.6191 0.7869
No log 2.02 202 0.6587 0.6203 0.6587 0.8116
No log 2.04 204 0.7177 0.5370 0.7177 0.8472
No log 2.06 206 0.6679 0.5794 0.6679 0.8172
No log 2.08 208 0.6305 0.6320 0.6305 0.7940
No log 2.1 210 0.8166 0.5411 0.8166 0.9037
No log 2.12 212 1.0385 0.4598 1.0385 1.0191
No log 2.14 214 1.0009 0.4782 1.0009 1.0005
No log 2.16 216 0.7879 0.5789 0.7879 0.8876
No log 2.18 218 0.6610 0.5603 0.6610 0.8130
No log 2.2 220 0.6457 0.6508 0.6457 0.8036
No log 2.22 222 0.6549 0.6970 0.6549 0.8093
No log 2.24 224 0.6354 0.6519 0.6354 0.7971
No log 2.26 226 0.6694 0.6487 0.6694 0.8182
No log 2.2800 228 0.6846 0.6318 0.6846 0.8274
No log 2.3 230 0.6368 0.6388 0.6368 0.7980
No log 2.32 232 0.6247 0.6306 0.6247 0.7904
No log 2.34 234 0.6257 0.6354 0.6257 0.7910
No log 2.36 236 0.6637 0.5459 0.6637 0.8147
No log 2.38 238 0.6733 0.5218 0.6733 0.8206
No log 2.4 240 0.6701 0.5383 0.6701 0.8186
No log 2.42 242 0.7303 0.5356 0.7303 0.8546
No log 2.44 244 0.7262 0.5356 0.7262 0.8522
No log 2.46 246 0.6899 0.6305 0.6899 0.8306
No log 2.48 248 0.6543 0.6041 0.6543 0.8089
No log 2.5 250 0.6990 0.6661 0.6990 0.8360
No log 2.52 252 0.7151 0.6751 0.7151 0.8457
No log 2.54 254 0.6897 0.6518 0.6897 0.8305
No log 2.56 256 0.7590 0.6171 0.7590 0.8712
No log 2.58 258 0.8100 0.5788 0.8100 0.9000
No log 2.6 260 0.7710 0.5993 0.7710 0.8780
No log 2.62 262 0.6857 0.6215 0.6857 0.8281
No log 2.64 264 0.6685 0.6095 0.6685 0.8176
No log 2.66 266 0.6642 0.7040 0.6642 0.8150
No log 2.68 268 0.6910 0.6341 0.6910 0.8312
No log 2.7 270 0.6650 0.6647 0.6650 0.8155
No log 2.7200 272 0.6472 0.7514 0.6472 0.8045
No log 2.74 274 0.7444 0.5993 0.7444 0.8628
No log 2.76 276 0.8091 0.5597 0.8091 0.8995
No log 2.7800 278 0.7848 0.5566 0.7848 0.8859
No log 2.8 280 0.6783 0.5959 0.6783 0.8236
No log 2.82 282 0.6421 0.5644 0.6421 0.8013
No log 2.84 284 0.6577 0.5879 0.6577 0.8110
No log 2.86 286 0.6531 0.6067 0.6531 0.8081
No log 2.88 288 0.6679 0.5632 0.6679 0.8173
No log 2.9 290 0.7761 0.5862 0.7761 0.8810
No log 2.92 292 0.8688 0.5799 0.8688 0.9321
No log 2.94 294 0.8222 0.5799 0.8222 0.9068
No log 2.96 296 0.7184 0.5818 0.7184 0.8476
No log 2.98 298 0.7015 0.6325 0.7015 0.8376
No log 3.0 300 0.6958 0.6615 0.6958 0.8341
No log 3.02 302 0.6712 0.5959 0.6712 0.8193
No log 3.04 304 0.6607 0.6111 0.6607 0.8128
No log 3.06 306 0.6796 0.6388 0.6796 0.8244
No log 3.08 308 0.6572 0.6388 0.6572 0.8107
No log 3.1 310 0.6162 0.6500 0.6162 0.7850
No log 3.12 312 0.6168 0.5939 0.6168 0.7854
No log 3.14 314 0.6182 0.5895 0.6182 0.7863
No log 3.16 316 0.6261 0.6078 0.6261 0.7913
No log 3.18 318 0.6368 0.5895 0.6368 0.7980
No log 3.2 320 0.6451 0.5919 0.6451 0.8032
No log 3.22 322 0.6403 0.6254 0.6403 0.8002
No log 3.24 324 0.6665 0.6066 0.6665 0.8164
No log 3.26 326 0.7236 0.6411 0.7236 0.8506
No log 3.2800 328 0.7317 0.6411 0.7317 0.8554
No log 3.3 330 0.6761 0.6083 0.6761 0.8223
No log 3.32 332 0.6293 0.6320 0.6293 0.7933
No log 3.34 334 0.6173 0.6114 0.6173 0.7857
No log 3.36 336 0.6225 0.6320 0.6225 0.7890
No log 3.38 338 0.6381 0.5621 0.6381 0.7988
No log 3.4 340 0.6300 0.6430 0.6300 0.7937
No log 3.42 342 0.6275 0.6206 0.6275 0.7921
No log 3.44 344 0.6311 0.6235 0.6311 0.7944
No log 3.46 346 0.6401 0.6085 0.6401 0.8001
No log 3.48 348 0.6669 0.5660 0.6669 0.8167
No log 3.5 350 0.6758 0.5660 0.6758 0.8221
No log 3.52 352 0.6897 0.5680 0.6897 0.8305
No log 3.54 354 0.6891 0.5249 0.6891 0.8301
No log 3.56 356 0.6930 0.5016 0.6930 0.8325
No log 3.58 358 0.7047 0.5345 0.7047 0.8395
No log 3.6 360 0.7245 0.4973 0.7245 0.8512
No log 3.62 362 0.7502 0.4697 0.7502 0.8662
No log 3.64 364 0.7804 0.4565 0.7804 0.8834
No log 3.66 366 0.8237 0.4568 0.8237 0.9076
No log 3.68 368 0.7960 0.5380 0.7960 0.8922
No log 3.7 370 0.7662 0.5972 0.7662 0.8753
No log 3.7200 372 0.7253 0.5977 0.7253 0.8517
No log 3.74 374 0.7712 0.5772 0.7712 0.8782
No log 3.76 376 0.7691 0.6054 0.7691 0.8770
No log 3.7800 378 0.7070 0.6127 0.7070 0.8408
No log 3.8 380 0.6711 0.6804 0.6711 0.8192
No log 3.82 382 0.6516 0.6679 0.6516 0.8072
No log 3.84 384 0.6462 0.6066 0.6462 0.8039
No log 3.86 386 0.6354 0.5552 0.6354 0.7971
No log 3.88 388 0.6207 0.5248 0.6207 0.7878
No log 3.9 390 0.6285 0.5820 0.6285 0.7928
No log 3.92 392 0.6259 0.6095 0.6259 0.7911
No log 3.94 394 0.6462 0.6087 0.6462 0.8039
No log 3.96 396 0.6874 0.6433 0.6874 0.8291
No log 3.98 398 0.6819 0.6203 0.6819 0.8258
No log 4.0 400 0.6840 0.6826 0.6840 0.8270
No log 4.02 402 0.7144 0.6664 0.7144 0.8452
No log 4.04 404 0.7246 0.5842 0.7246 0.8513
No log 4.06 406 0.7168 0.5954 0.7168 0.8467
No log 4.08 408 0.7107 0.5714 0.7107 0.8430
No log 4.1 410 0.7149 0.6087 0.7149 0.8455
No log 4.12 412 0.7158 0.6377 0.7158 0.8461
No log 4.14 414 0.7182 0.6826 0.7182 0.8475
No log 4.16 416 0.7518 0.6566 0.7518 0.8671
No log 4.18 418 0.7218 0.6630 0.7218 0.8496
No log 4.2 420 0.6589 0.6187 0.6589 0.8117
No log 4.22 422 0.6370 0.5905 0.6370 0.7981
No log 4.24 424 0.6316 0.5905 0.6316 0.7947
No log 4.26 426 0.6300 0.6399 0.6301 0.7938
No log 4.28 428 0.6529 0.6519 0.6529 0.8080
No log 4.3 430 0.6632 0.6721 0.6632 0.8144
No log 4.32 432 0.6346 0.6728 0.6346 0.7966
No log 4.34 434 0.6119 0.6439 0.6119 0.7823
No log 4.36 436 0.6098 0.6482 0.6098 0.7809
No log 4.38 438 0.6042 0.6781 0.6042 0.7773
No log 4.4 440 0.5975 0.6736 0.5975 0.7730
No log 4.42 442 0.5901 0.6570 0.5901 0.7682
No log 4.44 444 0.5807 0.6406 0.5807 0.7620
No log 4.46 446 0.5790 0.6320 0.5790 0.7609
No log 4.48 448 0.6047 0.6335 0.6047 0.7776
No log 4.5 450 0.6482 0.5912 0.6482 0.8051
No log 4.52 452 0.6709 0.5697 0.6709 0.8191
No log 4.54 454 0.6890 0.5584 0.6890 0.8301
No log 4.5600 456 0.6579 0.5666 0.6579 0.8111
No log 4.58 458 0.6385 0.5980 0.6385 0.7991
No log 4.6 460 0.6406 0.5964 0.6406 0.8004
No log 4.62 462 0.6402 0.6058 0.6402 0.8001
No log 4.64 464 0.6466 0.5599 0.6466 0.8041
No log 4.66 466 0.6628 0.5905 0.6628 0.8141
No log 4.68 468 0.6667 0.5376 0.6667 0.8165
No log 4.7 470 0.6610 0.5921 0.6610 0.8130
No log 4.72 472 0.6668 0.6374 0.6668 0.8166
No log 4.74 474 0.7228 0.6054 0.7228 0.8502
No log 4.76 476 0.7262 0.6071 0.7262 0.8522
No log 4.78 478 0.6842 0.5330 0.6842 0.8272
No log 4.8 480 0.6564 0.5174 0.6564 0.8102
No log 4.82 482 0.6568 0.5865 0.6568 0.8104
No log 4.84 484 0.6343 0.6350 0.6343 0.7964
No log 4.86 486 0.6364 0.6736 0.6364 0.7977
No log 4.88 488 0.6305 0.6864 0.6305 0.7940
No log 4.9 490 0.6658 0.6567 0.6658 0.8160
No log 4.92 492 0.6464 0.6559 0.6464 0.8040
No log 4.9400 494 0.6325 0.6360 0.6325 0.7953
No log 4.96 496 0.6510 0.6444 0.6510 0.8069
No log 4.98 498 0.7251 0.5584 0.7251 0.8515
0.3092 5.0 500 0.8071 0.5346 0.8071 0.8984
0.3092 5.02 502 0.8049 0.5021 0.8049 0.8972
0.3092 5.04 504 0.7738 0.5164 0.7738 0.8796
0.3092 5.06 506 0.7408 0.4824 0.7408 0.8607
0.3092 5.08 508 0.7061 0.5318 0.7061 0.8403
0.3092 5.1 510 0.7066 0.5697 0.7066 0.8406
0.3092 5.12 512 0.7506 0.5777 0.7506 0.8664
0.3092 5.14 514 0.7607 0.5924 0.7607 0.8722
0.3092 5.16 516 0.7076 0.6348 0.7076 0.8412
0.3092 5.18 518 0.6621 0.6272 0.6621 0.8137
0.3092 5.2 520 0.6701 0.5763 0.6701 0.8186
0.3092 5.22 522 0.6888 0.5752 0.6888 0.8300
0.3092 5.24 524 0.7343 0.5255 0.7343 0.8569
0.3092 5.26 526 0.7777 0.5033 0.7777 0.8819
0.3092 5.28 528 0.7482 0.4586 0.7482 0.8650
0.3092 5.3 530 0.7015 0.4227 0.7015 0.8375
0.3092 5.32 532 0.6901 0.4727 0.6901 0.8307

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k20_task5_organization

Finetuned
(4019)
this model