ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k15_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6849
  • Qwk: 0.5186
  • Mse: 0.6849
  • Rmse: 0.8276

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0267 2 3.7908 -0.0239 3.7908 1.9470
No log 0.0533 4 2.2226 0.1385 2.2226 1.4908
No log 0.08 6 1.1034 0.1711 1.1034 1.0504
No log 0.1067 8 1.1499 0.1471 1.1499 1.0724
No log 0.1333 10 1.0308 0.3815 1.0308 1.0153
No log 0.16 12 1.0314 0.0888 1.0314 1.0156
No log 0.1867 14 1.0444 0.2721 1.0444 1.0220
No log 0.2133 16 1.0026 0.2978 1.0026 1.0013
No log 0.24 18 0.9575 0.3048 0.9575 0.9785
No log 0.2667 20 0.9672 0.2787 0.9672 0.9835
No log 0.2933 22 0.9351 0.2687 0.9351 0.9670
No log 0.32 24 0.9078 0.3583 0.9078 0.9528
No log 0.3467 26 0.9101 0.3310 0.9101 0.9540
No log 0.3733 28 0.9150 0.3625 0.9150 0.9566
No log 0.4 30 0.8465 0.3625 0.8465 0.9201
No log 0.4267 32 0.8236 0.3813 0.8236 0.9075
No log 0.4533 34 0.8586 0.3328 0.8586 0.9266
No log 0.48 36 0.8431 0.3890 0.8431 0.9182
No log 0.5067 38 0.7020 0.4841 0.7020 0.8378
No log 0.5333 40 0.6632 0.5503 0.6632 0.8144
No log 0.56 42 0.6487 0.5862 0.6487 0.8054
No log 0.5867 44 0.6366 0.5932 0.6366 0.7979
No log 0.6133 46 0.6336 0.6500 0.6336 0.7960
No log 0.64 48 0.6448 0.6343 0.6448 0.8030
No log 0.6667 50 0.6275 0.6459 0.6275 0.7922
No log 0.6933 52 0.5381 0.6929 0.5381 0.7336
No log 0.72 54 0.5424 0.6356 0.5424 0.7365
No log 0.7467 56 0.5586 0.6626 0.5586 0.7474
No log 0.7733 58 0.6951 0.6845 0.6951 0.8337
No log 0.8 60 0.7257 0.6734 0.7257 0.8519
No log 0.8267 62 0.6337 0.6160 0.6337 0.7961
No log 0.8533 64 0.6380 0.5949 0.6380 0.7988
No log 0.88 66 0.6739 0.6986 0.6739 0.8209
No log 0.9067 68 0.6027 0.6774 0.6027 0.7763
No log 0.9333 70 0.5801 0.5763 0.5801 0.7617
No log 0.96 72 0.5765 0.5990 0.5765 0.7593
No log 0.9867 74 0.5798 0.5796 0.5798 0.7615
No log 1.0133 76 0.6264 0.6246 0.6264 0.7914
No log 1.04 78 0.6154 0.5985 0.6154 0.7845
No log 1.0667 80 0.6157 0.6032 0.6157 0.7847
No log 1.0933 82 0.6278 0.6032 0.6278 0.7924
No log 1.12 84 0.6372 0.6335 0.6372 0.7982
No log 1.1467 86 0.6617 0.6597 0.6617 0.8134
No log 1.1733 88 0.6705 0.6597 0.6705 0.8188
No log 1.2 90 0.6089 0.6881 0.6089 0.7803
No log 1.2267 92 0.6235 0.6408 0.6235 0.7896
No log 1.2533 94 0.6907 0.6311 0.6907 0.8311
No log 1.28 96 0.6733 0.6311 0.6733 0.8205
No log 1.3067 98 0.6003 0.6526 0.6003 0.7748
No log 1.3333 100 0.5798 0.6650 0.5798 0.7615
No log 1.3600 102 0.5798 0.6501 0.5798 0.7614
No log 1.3867 104 0.6143 0.6758 0.6143 0.7838
No log 1.4133 106 0.5969 0.7027 0.5969 0.7726
No log 1.44 108 0.5927 0.6377 0.5927 0.7699
No log 1.4667 110 0.6223 0.6410 0.6223 0.7889
No log 1.4933 112 0.6252 0.6703 0.6252 0.7907
No log 1.52 114 0.6385 0.6471 0.6385 0.7991
No log 1.5467 116 0.6338 0.6437 0.6338 0.7961
No log 1.5733 118 0.6392 0.5591 0.6392 0.7995
No log 1.6 120 0.6487 0.5671 0.6487 0.8054
No log 1.6267 122 0.7045 0.6266 0.7045 0.8394
No log 1.6533 124 0.7060 0.5763 0.7060 0.8402
No log 1.6800 126 0.6692 0.5833 0.6692 0.8181
No log 1.7067 128 0.7247 0.5586 0.7247 0.8513
No log 1.7333 130 0.7082 0.5455 0.7082 0.8416
No log 1.76 132 0.6494 0.5582 0.6494 0.8059
No log 1.7867 134 0.6509 0.5797 0.6509 0.8068
No log 1.8133 136 0.6378 0.6102 0.6378 0.7986
No log 1.8400 138 0.6149 0.6217 0.6149 0.7841
No log 1.8667 140 0.6333 0.6316 0.6333 0.7958
No log 1.8933 142 0.6288 0.6439 0.6288 0.7930
No log 1.92 144 0.7215 0.6198 0.7215 0.8494
No log 1.9467 146 0.9848 0.5069 0.9848 0.9924
No log 1.9733 148 0.9032 0.5076 0.9032 0.9503
No log 2.0 150 0.6603 0.7020 0.6603 0.8126
No log 2.0267 152 0.7527 0.5895 0.7527 0.8676
No log 2.0533 154 0.8304 0.5866 0.8304 0.9113
No log 2.08 156 0.6961 0.6448 0.6961 0.8343
No log 2.1067 158 0.6996 0.6578 0.6996 0.8364
No log 2.1333 160 0.9894 0.5213 0.9894 0.9947
No log 2.16 162 1.1217 0.4575 1.1217 1.0591
No log 2.1867 164 1.0097 0.5318 1.0097 1.0048
No log 2.2133 166 0.7725 0.5232 0.7725 0.8789
No log 2.24 168 0.6442 0.6133 0.6442 0.8026
No log 2.2667 170 0.6887 0.6072 0.6887 0.8299
No log 2.2933 172 0.6690 0.6919 0.6690 0.8179
No log 2.32 174 0.6295 0.6288 0.6295 0.7934
No log 2.3467 176 0.6833 0.5660 0.6833 0.8266
No log 2.3733 178 0.7785 0.5928 0.7785 0.8824
No log 2.4 180 0.7244 0.5844 0.7244 0.8511
No log 2.4267 182 0.5589 0.6348 0.5589 0.7476
No log 2.4533 184 0.5160 0.7110 0.5160 0.7183
No log 2.48 186 0.5611 0.6828 0.5611 0.7491
No log 2.5067 188 0.5390 0.7071 0.5390 0.7341
No log 2.5333 190 0.5456 0.6986 0.5456 0.7387
No log 2.56 192 0.6858 0.6690 0.6858 0.8281
No log 2.5867 194 0.9070 0.4770 0.9070 0.9524
No log 2.6133 196 0.8575 0.5378 0.8575 0.9260
No log 2.64 198 0.6595 0.6045 0.6595 0.8121
No log 2.6667 200 0.5748 0.6619 0.5748 0.7581
No log 2.6933 202 0.5810 0.6963 0.5810 0.7622
No log 2.7200 204 0.5556 0.6804 0.5556 0.7454
No log 2.7467 206 0.6012 0.6404 0.6012 0.7754
No log 2.7733 208 0.6726 0.5852 0.6726 0.8201
No log 2.8 210 0.7027 0.5854 0.7027 0.8383
No log 2.8267 212 0.6486 0.6228 0.6486 0.8053
No log 2.8533 214 0.6016 0.6197 0.6016 0.7757
No log 2.88 216 0.5899 0.6473 0.5899 0.7681
No log 2.9067 218 0.5834 0.6564 0.5834 0.7638
No log 2.9333 220 0.6016 0.6732 0.6016 0.7757
No log 2.96 222 0.7370 0.5715 0.7370 0.8585
No log 2.9867 224 0.8010 0.5495 0.8010 0.8950
No log 3.0133 226 0.7161 0.5686 0.7161 0.8462
No log 3.04 228 0.6229 0.5917 0.6229 0.7892
No log 3.0667 230 0.6330 0.6327 0.6330 0.7956
No log 3.0933 232 0.6295 0.6419 0.6295 0.7934
No log 3.12 234 0.6100 0.6390 0.6100 0.7810
No log 3.1467 236 0.6579 0.5968 0.6579 0.8111
No log 3.1733 238 0.7407 0.5715 0.7407 0.8607
No log 3.2 240 0.6778 0.6064 0.6778 0.8233
No log 3.2267 242 0.6240 0.6632 0.6240 0.7899
No log 3.2533 244 0.6551 0.6215 0.6551 0.8094
No log 3.2800 246 0.6627 0.6226 0.6627 0.8141
No log 3.3067 248 0.6374 0.6564 0.6374 0.7984
No log 3.3333 250 0.6623 0.6269 0.6623 0.8138
No log 3.36 252 0.6810 0.6296 0.6810 0.8252
No log 3.3867 254 0.6792 0.5529 0.6792 0.8241
No log 3.4133 256 0.6656 0.6164 0.6656 0.8159
No log 3.44 258 0.6862 0.6002 0.6862 0.8284
No log 3.4667 260 0.7458 0.5852 0.7458 0.8636
No log 3.4933 262 0.7803 0.6132 0.7803 0.8834
No log 3.52 264 0.7021 0.6318 0.7021 0.8379
No log 3.5467 266 0.6356 0.6581 0.6356 0.7972
No log 3.5733 268 0.6432 0.5934 0.6432 0.8020
No log 3.6 270 0.6491 0.6398 0.6491 0.8056
No log 3.6267 272 0.6776 0.6301 0.6776 0.8232
No log 3.6533 274 0.6702 0.6301 0.6702 0.8186
No log 3.68 276 0.6696 0.6460 0.6696 0.8183
No log 3.7067 278 0.6804 0.6464 0.6804 0.8249
No log 3.7333 280 0.6821 0.6740 0.6821 0.8259
No log 3.76 282 0.6959 0.6100 0.6959 0.8342
No log 3.7867 284 0.7094 0.5983 0.7094 0.8423
No log 3.8133 286 0.7781 0.5521 0.7781 0.8821
No log 3.84 288 0.7212 0.5902 0.7212 0.8493
No log 3.8667 290 0.6173 0.6092 0.6173 0.7857
No log 3.8933 292 0.5945 0.7056 0.5945 0.7710
No log 3.92 294 0.5902 0.7218 0.5902 0.7682
No log 3.9467 296 0.5976 0.6861 0.5976 0.7731
No log 3.9733 298 0.6198 0.6197 0.6198 0.7873
No log 4.0 300 0.6302 0.6084 0.6302 0.7939
No log 4.0267 302 0.6139 0.6697 0.6139 0.7835
No log 4.0533 304 0.6218 0.6812 0.6218 0.7885
No log 4.08 306 0.6041 0.6697 0.6041 0.7772
No log 4.1067 308 0.6294 0.6110 0.6294 0.7934
No log 4.1333 310 0.6882 0.5735 0.6882 0.8296
No log 4.16 312 0.7305 0.5526 0.7305 0.8547
No log 4.1867 314 0.6923 0.5019 0.6923 0.8320
No log 4.2133 316 0.6251 0.5917 0.6251 0.7906
No log 4.24 318 0.6051 0.6433 0.6051 0.7779
No log 4.2667 320 0.6226 0.6544 0.6226 0.7891
No log 4.2933 322 0.6138 0.7232 0.6138 0.7834
No log 4.32 324 0.6283 0.6609 0.6283 0.7927
No log 4.3467 326 0.7393 0.5604 0.7393 0.8598
No log 4.3733 328 0.7245 0.5428 0.7245 0.8512
No log 4.4 330 0.6477 0.5626 0.6477 0.8048
No log 4.4267 332 0.6340 0.6186 0.6340 0.7962
No log 4.4533 334 0.6353 0.6354 0.6353 0.7970
No log 4.48 336 0.6390 0.6916 0.6390 0.7994
No log 4.5067 338 0.6694 0.6322 0.6694 0.8182
No log 4.5333 340 0.6702 0.6578 0.6702 0.8187
No log 4.5600 342 0.6632 0.6717 0.6632 0.8144
No log 4.5867 344 0.6633 0.6553 0.6633 0.8144
No log 4.6133 346 0.6706 0.6447 0.6706 0.8189
No log 4.64 348 0.6967 0.6301 0.6967 0.8347
No log 4.6667 350 0.6943 0.6297 0.6943 0.8332
No log 4.6933 352 0.6966 0.6284 0.6966 0.8347
No log 4.72 354 0.7144 0.6240 0.7144 0.8452
No log 4.7467 356 0.7113 0.6046 0.7113 0.8434
No log 4.7733 358 0.7063 0.6486 0.7063 0.8404
No log 4.8 360 0.6857 0.6722 0.6857 0.8280
No log 4.8267 362 0.6719 0.6510 0.6719 0.8197
No log 4.8533 364 0.6801 0.5740 0.6801 0.8247
No log 4.88 366 0.7035 0.5059 0.7035 0.8387
No log 4.9067 368 0.6905 0.5175 0.6905 0.8310
No log 4.9333 370 0.6276 0.6177 0.6276 0.7922
No log 4.96 372 0.6209 0.6320 0.6209 0.7880
No log 4.9867 374 0.6365 0.5700 0.6365 0.7978
No log 5.0133 376 0.7204 0.5232 0.7204 0.8488
No log 5.04 378 0.7543 0.5232 0.7543 0.8685
No log 5.0667 380 0.7356 0.5318 0.7356 0.8577
No log 5.0933 382 0.6587 0.5380 0.6587 0.8116
No log 5.12 384 0.6167 0.6018 0.6167 0.7853
No log 5.1467 386 0.6102 0.6123 0.6102 0.7811
No log 5.1733 388 0.6153 0.6045 0.6153 0.7844
No log 5.2 390 0.6720 0.5842 0.6720 0.8197
No log 5.2267 392 0.7249 0.5821 0.7249 0.8514
No log 5.2533 394 0.6685 0.5902 0.6685 0.8176
No log 5.28 396 0.6089 0.5712 0.6089 0.7803
No log 5.3067 398 0.6296 0.5568 0.6296 0.7935
No log 5.3333 400 0.6218 0.5783 0.6218 0.7885
No log 5.36 402 0.6462 0.5429 0.6462 0.8039
No log 5.3867 404 0.7449 0.5675 0.7449 0.8631
No log 5.4133 406 0.7956 0.5714 0.7956 0.8920
No log 5.44 408 0.7261 0.5425 0.7261 0.8521
No log 5.4667 410 0.6349 0.5774 0.6349 0.7968
No log 5.4933 412 0.6162 0.6311 0.6162 0.7850
No log 5.52 414 0.6299 0.6985 0.6299 0.7937
No log 5.5467 416 0.6280 0.6874 0.6280 0.7924
No log 5.5733 418 0.6268 0.6473 0.6268 0.7917
No log 5.6 420 0.6923 0.6107 0.6923 0.8321
No log 5.6267 422 0.7705 0.5425 0.7705 0.8778
No log 5.6533 424 0.7403 0.5439 0.7403 0.8604
No log 5.68 426 0.7054 0.5291 0.7054 0.8399
No log 5.7067 428 0.6881 0.5373 0.6881 0.8295
No log 5.7333 430 0.6981 0.4643 0.6981 0.8355
No log 5.76 432 0.7030 0.5131 0.7030 0.8385
No log 5.7867 434 0.7218 0.5199 0.7218 0.8496
No log 5.8133 436 0.8180 0.5425 0.8180 0.9044
No log 5.84 438 0.8601 0.5102 0.8601 0.9274
No log 5.8667 440 0.7635 0.5144 0.7635 0.8738
No log 5.8933 442 0.6845 0.5375 0.6845 0.8274
No log 5.92 444 0.7298 0.4724 0.7298 0.8543
No log 5.9467 446 0.7640 0.4771 0.7640 0.8741
No log 5.9733 448 0.7079 0.4962 0.7079 0.8414
No log 6.0 450 0.6762 0.5359 0.6762 0.8223
No log 6.0267 452 0.7760 0.5342 0.7760 0.8809
No log 6.0533 454 0.7889 0.5232 0.7889 0.8882
No log 6.08 456 0.7056 0.5546 0.7056 0.8400
No log 6.1067 458 0.6691 0.4792 0.6691 0.8180
No log 6.1333 460 0.6808 0.4938 0.6808 0.8251
No log 6.16 462 0.6726 0.4643 0.6726 0.8201
No log 6.1867 464 0.6876 0.5546 0.6876 0.8292
No log 6.2133 466 0.7596 0.5318 0.7596 0.8715
No log 6.24 468 0.7416 0.5436 0.7416 0.8611
No log 6.2667 470 0.6652 0.5882 0.6652 0.8156
No log 6.2933 472 0.6245 0.6288 0.6245 0.7902
No log 6.32 474 0.6332 0.5882 0.6332 0.7958
No log 6.3467 476 0.6307 0.6007 0.6307 0.7942
No log 6.3733 478 0.6840 0.5279 0.6840 0.8270
No log 6.4 480 0.7758 0.5122 0.7758 0.8808
No log 6.4267 482 0.7819 0.5122 0.7819 0.8842
No log 6.4533 484 0.7032 0.5147 0.7032 0.8386
No log 6.48 486 0.6419 0.5637 0.6419 0.8012
No log 6.5067 488 0.6465 0.5522 0.6465 0.8041
No log 6.5333 490 0.6842 0.4836 0.6842 0.8272
No log 6.5600 492 0.6982 0.5046 0.6982 0.8356
No log 6.5867 494 0.6638 0.5186 0.6638 0.8148
No log 6.6133 496 0.6320 0.5438 0.6320 0.7950
No log 6.64 498 0.6446 0.5546 0.6446 0.8029
0.2757 6.6667 500 0.6626 0.5186 0.6626 0.8140
0.2757 6.6933 502 0.7326 0.5122 0.7326 0.8559
0.2757 6.72 504 0.7612 0.5428 0.7612 0.8725
0.2757 6.7467 506 0.7332 0.5025 0.7332 0.8563
0.2757 6.7733 508 0.6698 0.5199 0.6698 0.8184
0.2757 6.8 510 0.6456 0.5996 0.6456 0.8035
0.2757 6.8267 512 0.6454 0.6581 0.6454 0.8034
0.2757 6.8533 514 0.6520 0.6473 0.6520 0.8075
0.2757 6.88 516 0.6707 0.5657 0.6707 0.8190
0.2757 6.9067 518 0.7495 0.5025 0.7495 0.8657
0.2757 6.9333 520 0.7973 0.5006 0.7973 0.8929
0.2757 6.96 522 0.8148 0.5006 0.8148 0.9027
0.2757 6.9867 524 0.7665 0.5122 0.7665 0.8755
0.2757 7.0133 526 0.6849 0.5186 0.6849 0.8276

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k15_task5_organization

Finetuned
(4019)
this model