ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k5_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5809
  • Qwk: 0.6325
  • Mse: 0.5809
  • Rmse: 0.7622

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.08 2 3.8817 -0.0217 3.8817 1.9702
No log 0.16 4 2.0976 0.0727 2.0976 1.4483
No log 0.24 6 1.5549 -0.0417 1.5549 1.2469
No log 0.32 8 1.2587 0.0343 1.2587 1.1219
No log 0.4 10 1.1387 0.1284 1.1387 1.0671
No log 0.48 12 1.1908 0.1699 1.1908 1.0912
No log 0.56 14 1.1008 0.1935 1.1008 1.0492
No log 0.64 16 1.3175 0.0970 1.3175 1.1478
No log 0.72 18 1.6595 0.1975 1.6595 1.2882
No log 0.8 20 1.2825 0.2220 1.2825 1.1325
No log 0.88 22 0.8393 0.3730 0.8393 0.9161
No log 0.96 24 0.9591 0.2518 0.9591 0.9793
No log 1.04 26 0.9408 0.2518 0.9408 0.9699
No log 1.12 28 0.7815 0.3667 0.7815 0.8840
No log 1.2 30 0.8711 0.3066 0.8711 0.9333
No log 1.28 32 1.0114 0.2816 1.0114 1.0057
No log 1.3600 34 0.8554 0.3902 0.8554 0.9249
No log 1.44 36 0.6281 0.5887 0.6281 0.7925
No log 1.52 38 0.8109 0.6173 0.8109 0.9005
No log 1.6 40 0.7401 0.6709 0.7401 0.8603
No log 1.6800 42 0.6123 0.6058 0.6123 0.7825
No log 1.76 44 1.0412 0.4750 1.0412 1.0204
No log 1.8400 46 1.4529 0.3148 1.4529 1.2054
No log 1.92 48 1.1696 0.4500 1.1696 1.0815
No log 2.0 50 0.7799 0.5190 0.7799 0.8831
No log 2.08 52 0.7030 0.5964 0.7030 0.8384
No log 2.16 54 0.6848 0.5964 0.6848 0.8275
No log 2.24 56 0.8262 0.6283 0.8262 0.9089
No log 2.32 58 1.1590 0.4421 1.1590 1.0766
No log 2.4 60 1.1983 0.3737 1.1983 1.0946
No log 2.48 62 0.8460 0.5344 0.8460 0.9198
No log 2.56 64 0.6604 0.6254 0.6604 0.8127
No log 2.64 66 0.7355 0.6292 0.7355 0.8576
No log 2.7200 68 0.6632 0.6147 0.6632 0.8144
No log 2.8 70 0.6610 0.6246 0.6610 0.8130
No log 2.88 72 0.6591 0.6246 0.6591 0.8118
No log 2.96 74 0.6525 0.5955 0.6525 0.8078
No log 3.04 76 0.6890 0.6206 0.6890 0.8301
No log 3.12 78 0.6719 0.5931 0.6719 0.8197
No log 3.2 80 0.7909 0.6396 0.7909 0.8893
No log 3.2800 82 0.9247 0.5948 0.9247 0.9616
No log 3.36 84 1.0335 0.5264 1.0335 1.0166
No log 3.44 86 0.7811 0.6716 0.7811 0.8838
No log 3.52 88 0.6673 0.5892 0.6673 0.8169
No log 3.6 90 0.6643 0.5941 0.6643 0.8151
No log 3.68 92 0.7196 0.6108 0.7196 0.8483
No log 3.76 94 0.6957 0.6107 0.6957 0.8341
No log 3.84 96 0.6605 0.6634 0.6605 0.8127
No log 3.92 98 0.6512 0.6206 0.6512 0.8070
No log 4.0 100 0.7231 0.5659 0.7231 0.8504
No log 4.08 102 0.8106 0.5368 0.8106 0.9003
No log 4.16 104 0.7098 0.6411 0.7098 0.8425
No log 4.24 106 0.6851 0.5917 0.6851 0.8277
No log 4.32 108 0.7569 0.5865 0.7569 0.8700
No log 4.4 110 0.6795 0.5511 0.6795 0.8243
No log 4.48 112 0.7851 0.5792 0.7851 0.8861
No log 4.5600 114 1.1439 0.4077 1.1439 1.0695
No log 4.64 116 1.1376 0.4077 1.1376 1.0666
No log 4.72 118 0.8475 0.5157 0.8475 0.9206
No log 4.8 120 0.6410 0.5747 0.6410 0.8007
No log 4.88 122 0.6663 0.5968 0.6663 0.8163
No log 4.96 124 0.6590 0.6241 0.6590 0.8118
No log 5.04 126 0.6361 0.6386 0.6361 0.7976
No log 5.12 128 0.6566 0.6559 0.6566 0.8103
No log 5.2 130 0.6072 0.6499 0.6072 0.7792
No log 5.28 132 0.6098 0.6886 0.6098 0.7809
No log 5.36 134 0.6605 0.6071 0.6605 0.8127
No log 5.44 136 0.6693 0.6071 0.6693 0.8181
No log 5.52 138 0.6092 0.7081 0.6092 0.7805
No log 5.6 140 0.5944 0.6249 0.5944 0.7710
No log 5.68 142 0.6067 0.6195 0.6067 0.7789
No log 5.76 144 0.6435 0.6173 0.6435 0.8022
No log 5.84 146 0.6408 0.6354 0.6408 0.8005
No log 5.92 148 0.6495 0.6393 0.6495 0.8059
No log 6.0 150 0.6965 0.6346 0.6965 0.8346
No log 6.08 152 0.8483 0.5887 0.8483 0.9211
No log 6.16 154 0.7649 0.5239 0.7649 0.8746
No log 6.24 156 0.6514 0.6293 0.6514 0.8071
No log 6.32 158 0.6408 0.6407 0.6408 0.8005
No log 6.4 160 0.6596 0.5560 0.6596 0.8122
No log 6.48 162 0.7126 0.5447 0.7126 0.8442
No log 6.5600 164 0.7660 0.5855 0.7660 0.8752
No log 6.64 166 0.6336 0.6154 0.6336 0.7960
No log 6.72 168 0.6154 0.6620 0.6154 0.7845
No log 6.8 170 0.7117 0.5798 0.7117 0.8436
No log 6.88 172 0.6693 0.6198 0.6693 0.8181
No log 6.96 174 0.5991 0.6509 0.5991 0.7740
No log 7.04 176 0.6303 0.6526 0.6303 0.7939
No log 7.12 178 0.6439 0.5865 0.6439 0.8024
No log 7.2 180 0.6920 0.5870 0.6920 0.8318
No log 7.28 182 0.7052 0.6174 0.7052 0.8398
No log 7.36 184 0.6127 0.6804 0.6127 0.7827
No log 7.44 186 0.5901 0.6272 0.5901 0.7682
No log 7.52 188 0.5840 0.6501 0.5840 0.7642
No log 7.6 190 0.5742 0.6882 0.5742 0.7577
No log 7.68 192 0.5912 0.6759 0.5912 0.7689
No log 7.76 194 0.5837 0.6427 0.5837 0.7640
No log 7.84 196 0.5978 0.6265 0.5978 0.7732
No log 7.92 198 0.6049 0.6377 0.6049 0.7778
No log 8.0 200 0.5837 0.6109 0.5837 0.7640
No log 8.08 202 0.5528 0.6699 0.5528 0.7435
No log 8.16 204 0.6184 0.5860 0.6184 0.7864
No log 8.24 206 0.6550 0.6283 0.6550 0.8093
No log 8.32 208 0.6592 0.6312 0.6592 0.8119
No log 8.4 210 0.5578 0.6778 0.5578 0.7468
No log 8.48 212 0.5570 0.6415 0.5570 0.7464
No log 8.56 214 0.5729 0.6317 0.5729 0.7569
No log 8.64 216 0.5870 0.5894 0.5870 0.7662
No log 8.72 218 0.5653 0.6108 0.5653 0.7518
No log 8.8 220 0.5286 0.6507 0.5286 0.7270
No log 8.88 222 0.5374 0.6929 0.5374 0.7331
No log 8.96 224 0.5471 0.6245 0.5471 0.7396
No log 9.04 226 0.4932 0.7088 0.4932 0.7023
No log 9.12 228 0.4997 0.7064 0.4997 0.7069
No log 9.2 230 0.5548 0.6880 0.5548 0.7448
No log 9.28 232 0.6011 0.6664 0.6011 0.7753
No log 9.36 234 0.6275 0.6312 0.6275 0.7922
No log 9.44 236 0.5303 0.6733 0.5303 0.7282
No log 9.52 238 0.5039 0.6164 0.5039 0.7099
No log 9.6 240 0.5579 0.6678 0.5579 0.7469
No log 9.68 242 0.5385 0.7114 0.5385 0.7338
No log 9.76 244 0.4918 0.6546 0.4918 0.7013
No log 9.84 246 0.5039 0.6935 0.5039 0.7098
No log 9.92 248 0.5059 0.7049 0.5059 0.7113
No log 10.0 250 0.4972 0.6488 0.4972 0.7051
No log 10.08 252 0.5092 0.6716 0.5092 0.7136
No log 10.16 254 0.5141 0.6725 0.5141 0.7170
No log 10.24 256 0.6542 0.6610 0.6542 0.8088
No log 10.32 258 0.6912 0.6706 0.6912 0.8314
No log 10.4 260 0.5498 0.6964 0.5498 0.7415
No log 10.48 262 0.5038 0.7082 0.5038 0.7098
No log 10.56 264 0.5307 0.6648 0.5307 0.7285
No log 10.64 266 0.4985 0.6980 0.4985 0.7061
No log 10.72 268 0.5232 0.6676 0.5232 0.7233
No log 10.8 270 0.5649 0.6172 0.5649 0.7516
No log 10.88 272 0.5701 0.5692 0.5701 0.7551
No log 10.96 274 0.5754 0.5874 0.5754 0.7586
No log 11.04 276 0.6347 0.5825 0.6347 0.7967
No log 11.12 278 0.6812 0.6190 0.6812 0.8253
No log 11.2 280 0.6486 0.5877 0.6486 0.8054
No log 11.28 282 0.5867 0.6511 0.5867 0.7659
No log 11.36 284 0.5502 0.6389 0.5502 0.7417
No log 11.44 286 0.5470 0.6822 0.5470 0.7396
No log 11.52 288 0.5896 0.6332 0.5896 0.7679
No log 11.6 290 0.5830 0.6332 0.5830 0.7635
No log 11.68 292 0.5433 0.6676 0.5433 0.7371
No log 11.76 294 0.5253 0.6460 0.5253 0.7248
No log 11.84 296 0.5075 0.6584 0.5075 0.7124
No log 11.92 298 0.5088 0.7001 0.5088 0.7133
No log 12.0 300 0.5184 0.6590 0.5184 0.7200
No log 12.08 302 0.4813 0.7323 0.4813 0.6938
No log 12.16 304 0.4792 0.7077 0.4792 0.6922
No log 12.24 306 0.5365 0.6768 0.5365 0.7324
No log 12.32 308 0.5065 0.6894 0.5065 0.7117
No log 12.4 310 0.4818 0.6830 0.4818 0.6941
No log 12.48 312 0.4885 0.6364 0.4885 0.6989
No log 12.56 314 0.5019 0.6456 0.5019 0.7084
No log 12.64 316 0.5163 0.6652 0.5163 0.7186
No log 12.72 318 0.5161 0.6652 0.5161 0.7184
No log 12.8 320 0.5039 0.6249 0.5039 0.7099
No log 12.88 322 0.5650 0.6807 0.5650 0.7516
No log 12.96 324 0.7567 0.6190 0.7567 0.8699
No log 13.04 326 0.8491 0.5952 0.8491 0.9215
No log 13.12 328 0.7223 0.6209 0.7223 0.8499
No log 13.2 330 0.5357 0.6748 0.5357 0.7319
No log 13.28 332 0.4889 0.6898 0.4889 0.6992
No log 13.36 334 0.5351 0.7202 0.5351 0.7315
No log 13.44 336 0.5176 0.7040 0.5176 0.7194
No log 13.52 338 0.4985 0.6716 0.4985 0.7061
No log 13.6 340 0.5187 0.6557 0.5187 0.7202
No log 13.68 342 0.5254 0.6724 0.5254 0.7248
No log 13.76 344 0.5580 0.6639 0.5580 0.7470
No log 13.84 346 0.5158 0.6664 0.5158 0.7182
No log 13.92 348 0.4990 0.7049 0.4990 0.7064
No log 14.0 350 0.5360 0.7040 0.5360 0.7321
No log 14.08 352 0.5290 0.6943 0.5290 0.7274
No log 14.16 354 0.5009 0.6207 0.5009 0.7077
No log 14.24 356 0.5234 0.6451 0.5234 0.7235
No log 14.32 358 0.5613 0.6748 0.5613 0.7492
No log 14.4 360 0.5459 0.6748 0.5459 0.7389
No log 14.48 362 0.5129 0.6354 0.5129 0.7162
No log 14.56 364 0.5145 0.6207 0.5145 0.7173
No log 14.64 366 0.5188 0.6107 0.5188 0.7203
No log 14.72 368 0.5183 0.6353 0.5183 0.7199
No log 14.8 370 0.5259 0.6353 0.5259 0.7252
No log 14.88 372 0.5225 0.6272 0.5225 0.7228
No log 14.96 374 0.5111 0.6649 0.5111 0.7149
No log 15.04 376 0.5130 0.7005 0.5130 0.7162
No log 15.12 378 0.5344 0.6830 0.5344 0.7310
No log 15.2 380 0.5456 0.6830 0.5456 0.7387
No log 15.28 382 0.5404 0.6641 0.5404 0.7351
No log 15.36 384 0.5429 0.6164 0.5429 0.7368
No log 15.44 386 0.5512 0.6230 0.5512 0.7424
No log 15.52 388 0.5485 0.6164 0.5485 0.7406
No log 15.6 390 0.5420 0.6354 0.5420 0.7362
No log 15.68 392 0.5428 0.6354 0.5428 0.7367
No log 15.76 394 0.5582 0.6230 0.5582 0.7471
No log 15.84 396 0.5825 0.6343 0.5825 0.7632
No log 15.92 398 0.5936 0.6209 0.5936 0.7705
No log 16.0 400 0.6282 0.5763 0.6282 0.7926
No log 16.08 402 0.6256 0.5763 0.6256 0.7909
No log 16.16 404 0.5682 0.6508 0.5682 0.7538
No log 16.24 406 0.5236 0.6742 0.5236 0.7236
No log 16.32 408 0.5538 0.6985 0.5538 0.7441
No log 16.4 410 0.5800 0.6578 0.5800 0.7616
No log 16.48 412 0.5334 0.7145 0.5334 0.7303
No log 16.56 414 0.5072 0.6634 0.5072 0.7122
No log 16.64 416 0.5226 0.6435 0.5226 0.7229
No log 16.72 418 0.5360 0.6276 0.5360 0.7321
No log 16.8 420 0.5083 0.6186 0.5083 0.7129
No log 16.88 422 0.5034 0.6753 0.5034 0.7095
No log 16.96 424 0.5132 0.6952 0.5132 0.7164
No log 17.04 426 0.5127 0.7104 0.5127 0.7161
No log 17.12 428 0.5157 0.7104 0.5157 0.7181
No log 17.2 430 0.5543 0.7047 0.5543 0.7445
No log 17.28 432 0.5637 0.7047 0.5637 0.7508
No log 17.36 434 0.5300 0.6497 0.5300 0.7280
No log 17.44 436 0.5799 0.6084 0.5799 0.7615
No log 17.52 438 0.7208 0.6521 0.7208 0.8490
No log 17.6 440 0.8163 0.6229 0.8163 0.9035
No log 17.68 442 0.7951 0.6446 0.7951 0.8917
No log 17.76 444 0.6853 0.6227 0.6853 0.8278
No log 17.84 446 0.5800 0.6324 0.5800 0.7615
No log 17.92 448 0.5461 0.6564 0.5461 0.7390
No log 18.0 450 0.5928 0.6402 0.5928 0.7699
No log 18.08 452 0.6100 0.6308 0.6100 0.7811
No log 18.16 454 0.5633 0.6507 0.5633 0.7505
No log 18.24 456 0.5329 0.6442 0.5329 0.7300
No log 18.32 458 0.5927 0.7013 0.5927 0.7699
No log 18.4 460 0.6553 0.6864 0.6553 0.8095
No log 18.48 462 0.6425 0.6906 0.6425 0.8016
No log 18.56 464 0.5790 0.6609 0.5790 0.7610
No log 18.64 466 0.5333 0.6627 0.5333 0.7303
No log 18.72 468 0.5823 0.6692 0.5823 0.7631
No log 18.8 470 0.6101 0.6603 0.6101 0.7811
No log 18.88 472 0.5794 0.6595 0.5794 0.7612
No log 18.96 474 0.5529 0.7096 0.5529 0.7436
No log 19.04 476 0.5577 0.7219 0.5577 0.7468
No log 19.12 478 0.5676 0.7063 0.5676 0.7534
No log 19.2 480 0.5627 0.6901 0.5627 0.7501
No log 19.28 482 0.5649 0.6297 0.5649 0.7516
No log 19.36 484 0.5700 0.6039 0.5700 0.7550
No log 19.44 486 0.5782 0.6097 0.5782 0.7604
No log 19.52 488 0.5765 0.6097 0.5765 0.7593
No log 19.6 490 0.5826 0.6097 0.5826 0.7633
No log 19.68 492 0.5813 0.6097 0.5813 0.7624
No log 19.76 494 0.5765 0.6507 0.5765 0.7593
No log 19.84 496 0.5709 0.6606 0.5709 0.7556
No log 19.92 498 0.5495 0.7034 0.5495 0.7413
0.2411 20.0 500 0.5451 0.6709 0.5451 0.7383
0.2411 20.08 502 0.5449 0.6805 0.5449 0.7382
0.2411 20.16 504 0.5442 0.6788 0.5442 0.7377
0.2411 20.24 506 0.5572 0.6196 0.5572 0.7465
0.2411 20.32 508 0.5794 0.6024 0.5794 0.7612
0.2411 20.4 510 0.5895 0.6139 0.5895 0.7678
0.2411 20.48 512 0.5868 0.6232 0.5868 0.7660
0.2411 20.56 514 0.5809 0.6325 0.5809 0.7622

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k5_task5_organization

Finetuned
(4019)
this model