ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k7_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7112
  • Qwk: 0.5267
  • Mse: 0.7112
  • Rmse: 0.8433

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0571 2 3.8326 -0.0347 3.8326 1.9577
No log 0.1143 4 1.9256 -0.0082 1.9256 1.3877
No log 0.1714 6 1.2295 -0.0064 1.2295 1.1088
No log 0.2286 8 0.9938 0.2108 0.9938 0.9969
No log 0.2857 10 1.0113 0.1389 1.0113 1.0057
No log 0.3429 12 0.9739 0.1908 0.9739 0.9868
No log 0.4 14 1.0065 0.2981 1.0065 1.0032
No log 0.4571 16 1.1214 0.1790 1.1214 1.0589
No log 0.5143 18 1.1363 0.2002 1.1363 1.0660
No log 0.5714 20 0.9994 0.2659 0.9994 0.9997
No log 0.6286 22 0.9098 0.2770 0.9098 0.9539
No log 0.6857 24 1.1137 0.0888 1.1137 1.0553
No log 0.7429 26 1.2171 0.2424 1.2171 1.1032
No log 0.8 28 1.0303 0.3381 1.0303 1.0150
No log 0.8571 30 0.8543 0.5346 0.8543 0.9243
No log 0.9143 32 0.7865 0.4691 0.7865 0.8869
No log 0.9714 34 0.8862 0.4299 0.8862 0.9414
No log 1.0286 36 0.8860 0.3812 0.8860 0.9413
No log 1.0857 38 0.7734 0.4757 0.7734 0.8794
No log 1.1429 40 0.6959 0.5469 0.6959 0.8342
No log 1.2 42 0.7542 0.4067 0.7542 0.8685
No log 1.2571 44 0.7976 0.3922 0.7976 0.8931
No log 1.3143 46 0.6951 0.5057 0.6951 0.8337
No log 1.3714 48 0.6386 0.5902 0.6386 0.7991
No log 1.4286 50 0.5987 0.5759 0.5987 0.7738
No log 1.4857 52 0.6044 0.5331 0.6044 0.7774
No log 1.5429 54 0.6014 0.6122 0.6014 0.7755
No log 1.6 56 0.7392 0.5833 0.7392 0.8598
No log 1.6571 58 0.8070 0.5607 0.8070 0.8983
No log 1.7143 60 0.6747 0.6670 0.6747 0.8214
No log 1.7714 62 0.5877 0.4784 0.5877 0.7666
No log 1.8286 64 0.7417 0.6362 0.7417 0.8612
No log 1.8857 66 0.7272 0.6362 0.7272 0.8528
No log 1.9429 68 0.6161 0.5796 0.6161 0.7849
No log 2.0 70 0.6785 0.5898 0.6785 0.8237
No log 2.0571 72 0.7229 0.6278 0.7229 0.8502
No log 2.1143 74 0.6648 0.5832 0.6648 0.8153
No log 2.1714 76 0.6618 0.5288 0.6618 0.8135
No log 2.2286 78 0.7528 0.5793 0.7528 0.8677
No log 2.2857 80 0.7407 0.5688 0.7407 0.8607
No log 2.3429 82 0.6530 0.5635 0.6530 0.8081
No log 2.4 84 0.6554 0.4982 0.6554 0.8095
No log 2.4571 86 0.6716 0.5973 0.6716 0.8195
No log 2.5143 88 0.6244 0.5546 0.6244 0.7902
No log 2.5714 90 0.6416 0.6420 0.6416 0.8010
No log 2.6286 92 0.6359 0.5909 0.6359 0.7974
No log 2.6857 94 0.6268 0.5832 0.6268 0.7917
No log 2.7429 96 0.6272 0.6558 0.6272 0.7919
No log 2.8 98 0.5834 0.5690 0.5834 0.7638
No log 2.8571 100 0.6142 0.5879 0.6142 0.7837
No log 2.9143 102 0.6559 0.5982 0.6559 0.8099
No log 2.9714 104 0.6009 0.5966 0.6009 0.7752
No log 3.0286 106 0.6298 0.6319 0.6298 0.7936
No log 3.0857 108 0.6807 0.5998 0.6807 0.8250
No log 3.1429 110 0.6544 0.5798 0.6544 0.8089
No log 3.2 112 0.6056 0.6333 0.6056 0.7782
No log 3.2571 114 0.6376 0.6544 0.6376 0.7985
No log 3.3143 116 0.6544 0.6435 0.6544 0.8090
No log 3.3714 118 0.6270 0.6269 0.6270 0.7918
No log 3.4286 120 0.6306 0.6259 0.6306 0.7941
No log 3.4857 122 0.7311 0.5777 0.7311 0.8550
No log 3.5429 124 0.7157 0.5584 0.7157 0.8460
No log 3.6 126 0.6597 0.5395 0.6597 0.8122
No log 3.6571 128 0.6394 0.6164 0.6394 0.7996
No log 3.7143 130 0.6756 0.5554 0.6756 0.8219
No log 3.7714 132 0.6590 0.5830 0.6590 0.8118
No log 3.8286 134 0.6582 0.5763 0.6582 0.8113
No log 3.8857 136 0.6489 0.5861 0.6489 0.8055
No log 3.9429 138 0.6226 0.5991 0.6226 0.7890
No log 4.0 140 0.6007 0.6215 0.6007 0.7751
No log 4.0571 142 0.6172 0.5993 0.6172 0.7856
No log 4.1143 144 0.6213 0.5676 0.6213 0.7882
No log 4.1714 146 0.5838 0.6626 0.5838 0.7641
No log 4.2286 148 0.7192 0.5696 0.7192 0.8481
No log 4.2857 150 0.7053 0.5425 0.7053 0.8398
No log 4.3429 152 0.6034 0.6209 0.6034 0.7768
No log 4.4 154 0.6232 0.5999 0.6232 0.7894
No log 4.4571 156 0.6213 0.6006 0.6213 0.7882
No log 4.5143 158 0.6383 0.6133 0.6383 0.7989
No log 4.5714 160 0.7882 0.5505 0.7882 0.8878
No log 4.6286 162 0.8435 0.5387 0.8435 0.9184
No log 4.6857 164 0.7938 0.5715 0.7938 0.8909
No log 4.7429 166 0.7296 0.5543 0.7296 0.8542
No log 4.8 168 0.6536 0.5523 0.6536 0.8085
No log 4.8571 170 0.6700 0.5882 0.6700 0.8185
No log 4.9143 172 0.6647 0.5898 0.6647 0.8153
No log 4.9714 174 0.6799 0.5288 0.6799 0.8246
No log 5.0286 176 0.7270 0.4713 0.7270 0.8526
No log 5.0857 178 0.7689 0.4597 0.7689 0.8769
No log 5.1429 180 0.7651 0.4601 0.7651 0.8747
No log 5.2 182 0.6654 0.5442 0.6654 0.8157
No log 5.2571 184 0.6268 0.5383 0.6268 0.7917
No log 5.3143 186 0.6333 0.5934 0.6333 0.7958
No log 5.3714 188 0.6367 0.6306 0.6367 0.7979
No log 5.4286 190 0.6457 0.6089 0.6457 0.8036
No log 5.4857 192 0.6130 0.6383 0.6130 0.7829
No log 5.5429 194 0.6141 0.5798 0.6141 0.7837
No log 5.6 196 0.6720 0.5232 0.6720 0.8198
No log 5.6571 198 0.6497 0.5663 0.6497 0.8060
No log 5.7143 200 0.6047 0.5698 0.6047 0.7776
No log 5.7714 202 0.5867 0.6143 0.5867 0.7660
No log 5.8286 204 0.6355 0.5746 0.6355 0.7972
No log 5.8857 206 0.7073 0.6102 0.7073 0.8410
No log 5.9429 208 0.6268 0.6471 0.6268 0.7917
No log 6.0 210 0.6087 0.6289 0.6087 0.7802
No log 6.0571 212 0.5820 0.6242 0.5820 0.7629
No log 6.1143 214 0.6158 0.5912 0.6158 0.7848
No log 6.1714 216 0.5732 0.6944 0.5732 0.7571
No log 6.2286 218 0.6626 0.5809 0.6626 0.8140
No log 6.2857 220 0.6874 0.5631 0.6874 0.8291
No log 6.3429 222 0.6372 0.5688 0.6372 0.7982
No log 6.4 224 0.6760 0.5510 0.6760 0.8222
No log 6.4571 226 0.6942 0.5498 0.6942 0.8332
No log 6.5143 228 0.6382 0.5787 0.6382 0.7989
No log 6.5714 230 0.6026 0.6282 0.6026 0.7763
No log 6.6286 232 0.5913 0.6282 0.5913 0.7690
No log 6.6857 234 0.6311 0.5699 0.6311 0.7944
No log 6.7429 236 0.6587 0.5746 0.6587 0.8116
No log 6.8 238 0.5984 0.6066 0.5984 0.7736
No log 6.8571 240 0.5903 0.5843 0.5903 0.7683
No log 6.9143 242 0.5839 0.6498 0.5839 0.7641
No log 6.9714 244 0.5881 0.6606 0.5881 0.7669
No log 7.0286 246 0.6005 0.6133 0.6005 0.7749
No log 7.0857 248 0.6051 0.6597 0.6051 0.7779
No log 7.1429 250 0.6322 0.5407 0.6322 0.7951
No log 7.2 252 0.6197 0.5407 0.6197 0.7872
No log 7.2571 254 0.5986 0.5977 0.5986 0.7737
No log 7.3143 256 0.6065 0.5650 0.6065 0.7788
No log 7.3714 258 0.5919 0.6470 0.5919 0.7694
No log 7.4286 260 0.6030 0.5830 0.6030 0.7765
No log 7.4857 262 0.6234 0.5923 0.6234 0.7895
No log 7.5429 264 0.6478 0.5810 0.6478 0.8048
No log 7.6 266 0.6190 0.5438 0.6190 0.7867
No log 7.6571 268 0.6322 0.4947 0.6322 0.7951
No log 7.7143 270 0.6225 0.4893 0.6225 0.7890
No log 7.7714 272 0.6911 0.5799 0.6911 0.8313
No log 7.8286 274 0.7480 0.5647 0.7480 0.8649
No log 7.8857 276 0.6878 0.5266 0.6878 0.8293
No log 7.9429 278 0.6657 0.5572 0.6657 0.8159
No log 8.0 280 0.6438 0.5597 0.6438 0.8024
No log 8.0571 282 0.6223 0.5724 0.6223 0.7889
No log 8.1143 284 0.6226 0.5809 0.6226 0.7891
No log 8.1714 286 0.6356 0.5990 0.6356 0.7972
No log 8.2286 288 0.6513 0.5877 0.6513 0.8070
No log 8.2857 290 0.6399 0.5887 0.6399 0.7999
No log 8.3429 292 0.6547 0.5563 0.6547 0.8091
No log 8.4 294 0.6656 0.5346 0.6656 0.8158
No log 8.4571 296 0.6702 0.4873 0.6702 0.8186
No log 8.5143 298 0.7090 0.5521 0.7090 0.8420
No log 8.5714 300 0.6860 0.5059 0.6860 0.8283
No log 8.6286 302 0.6337 0.5279 0.6337 0.7961
No log 8.6857 304 0.6155 0.5766 0.6155 0.7845
No log 8.7429 306 0.6507 0.5466 0.6507 0.8067
No log 8.8 308 0.7704 0.5811 0.7704 0.8777
No log 8.8571 310 0.7562 0.5431 0.7562 0.8696
No log 8.9143 312 0.6811 0.5107 0.6811 0.8253
No log 8.9714 314 0.6708 0.5174 0.6708 0.8190
No log 9.0286 316 0.7042 0.5548 0.7042 0.8392
No log 9.0857 318 0.6867 0.5548 0.6867 0.8287
No log 9.1429 320 0.6436 0.5176 0.6436 0.8023
No log 9.2 322 0.6889 0.5810 0.6889 0.8300
No log 9.2571 324 0.7689 0.5655 0.7689 0.8769
No log 9.3143 326 0.7458 0.5343 0.7458 0.8636
No log 9.3714 328 0.6407 0.5098 0.6407 0.8005
No log 9.4286 330 0.6670 0.6205 0.6670 0.8167
No log 9.4857 332 0.7047 0.6032 0.7047 0.8395
No log 9.5429 334 0.6654 0.5174 0.6654 0.8157
No log 9.6 336 0.7201 0.4727 0.7201 0.8486
No log 9.6571 338 0.8905 0.5695 0.8905 0.9437
No log 9.7143 340 0.8919 0.5740 0.8919 0.9444
No log 9.7714 342 0.7461 0.5628 0.7461 0.8638
No log 9.8286 344 0.6366 0.5356 0.6366 0.7979
No log 9.8857 346 0.6229 0.5923 0.6229 0.7892
No log 9.9429 348 0.6308 0.5236 0.6308 0.7942
No log 10.0 350 0.6825 0.5912 0.6825 0.8262
No log 10.0571 352 0.7329 0.5666 0.7329 0.8561
No log 10.1143 354 0.7365 0.5677 0.7365 0.8582
No log 10.1714 356 0.7130 0.5175 0.7130 0.8444
No log 10.2286 358 0.7055 0.5459 0.7055 0.8399
No log 10.2857 360 0.6985 0.5107 0.6985 0.8358
No log 10.3429 362 0.6992 0.4873 0.6992 0.8362
No log 10.4 364 0.7012 0.4378 0.7012 0.8374
No log 10.4571 366 0.7291 0.5446 0.7291 0.8539
No log 10.5143 368 0.7623 0.5279 0.7623 0.8731
No log 10.5714 370 0.7409 0.5475 0.7409 0.8607
No log 10.6286 372 0.6846 0.5614 0.6846 0.8274
No log 10.6857 374 0.6748 0.5489 0.6748 0.8214
No log 10.7429 376 0.6588 0.5808 0.6588 0.8117
No log 10.8 378 0.6742 0.5808 0.6742 0.8211
No log 10.8571 380 0.7115 0.5455 0.7115 0.8435
No log 10.9143 382 0.6732 0.5869 0.6732 0.8205
No log 10.9714 384 0.6326 0.5142 0.6326 0.7954
No log 11.0286 386 0.6227 0.5945 0.6227 0.7891
No log 11.0857 388 0.6195 0.6039 0.6195 0.7871
No log 11.1429 390 0.6255 0.5831 0.6255 0.7909
No log 11.2 392 0.6414 0.5133 0.6414 0.8009
No log 11.2571 394 0.6566 0.5751 0.6566 0.8103
No log 11.3143 396 0.6406 0.5719 0.6406 0.8004
No log 11.3714 398 0.6150 0.5980 0.6150 0.7842
No log 11.4286 400 0.6196 0.5980 0.6196 0.7871
No log 11.4857 402 0.6046 0.5557 0.6046 0.7776
No log 11.5429 404 0.6120 0.5660 0.6120 0.7823
No log 11.6 406 0.6171 0.5648 0.6171 0.7856
No log 11.6571 408 0.5973 0.5808 0.5973 0.7728
No log 11.7143 410 0.5915 0.5854 0.5915 0.7691
No log 11.7714 412 0.5852 0.6124 0.5852 0.7650
No log 11.8286 414 0.5794 0.6456 0.5794 0.7612
No log 11.8857 416 0.5977 0.6254 0.5977 0.7731
No log 11.9429 418 0.6508 0.6362 0.6508 0.8068
No log 12.0 420 0.6898 0.6123 0.6898 0.8306
No log 12.0571 422 0.6356 0.5885 0.6356 0.7972
No log 12.1143 424 0.6075 0.5923 0.6075 0.7794
No log 12.1714 426 0.6120 0.5713 0.6120 0.7823
No log 12.2286 428 0.6127 0.5594 0.6127 0.7827
No log 12.2857 430 0.6084 0.5796 0.6084 0.7800
No log 12.3429 432 0.5966 0.5774 0.5966 0.7724
No log 12.4 434 0.6035 0.6388 0.6035 0.7769
No log 12.4571 436 0.6147 0.6485 0.6147 0.7840
No log 12.5143 438 0.5905 0.6001 0.5905 0.7684
No log 12.5714 440 0.5861 0.6144 0.5861 0.7656
No log 12.6286 442 0.5849 0.6028 0.5849 0.7648
No log 12.6857 444 0.5712 0.6345 0.5712 0.7558
No log 12.7429 446 0.5696 0.6641 0.5696 0.7547
No log 12.8 448 0.5717 0.6988 0.5717 0.7561
No log 12.8571 450 0.5734 0.6944 0.5734 0.7572
No log 12.9143 452 0.5702 0.6720 0.5702 0.7551
No log 12.9714 454 0.5675 0.6720 0.5675 0.7533
No log 13.0286 456 0.5790 0.6491 0.5790 0.7609
No log 13.0857 458 0.5917 0.6589 0.5917 0.7692
No log 13.1429 460 0.5917 0.6446 0.5917 0.7692
No log 13.2 462 0.5740 0.6491 0.5740 0.7577
No log 13.2571 464 0.5662 0.7034 0.5662 0.7525
No log 13.3143 466 0.5636 0.6339 0.5636 0.7507
No log 13.3714 468 0.5712 0.6414 0.5712 0.7558
No log 13.4286 470 0.5922 0.6555 0.5922 0.7695
No log 13.4857 472 0.6207 0.6018 0.6207 0.7878
No log 13.5429 474 0.6459 0.5923 0.6459 0.8037
No log 13.6 476 0.6529 0.5699 0.6529 0.8080
No log 13.6571 478 0.6431 0.5977 0.6431 0.8019
No log 13.7143 480 0.6467 0.6197 0.6467 0.8042
No log 13.7714 482 0.6394 0.6353 0.6394 0.7996
No log 13.8286 484 0.6392 0.6353 0.6392 0.7995
No log 13.8857 486 0.6214 0.6568 0.6214 0.7883
No log 13.9429 488 0.5978 0.6545 0.5978 0.7732
No log 14.0 490 0.6113 0.6339 0.6113 0.7818
No log 14.0571 492 0.6124 0.6339 0.6124 0.7826
No log 14.1143 494 0.6045 0.6500 0.6045 0.7775
No log 14.1714 496 0.6411 0.6035 0.6411 0.8007
No log 14.2286 498 0.6797 0.5477 0.6797 0.8244
0.2521 14.2857 500 0.6398 0.6207 0.6398 0.7999
0.2521 14.3429 502 0.5991 0.6680 0.5991 0.7740
0.2521 14.4 504 0.6117 0.6067 0.6117 0.7821
0.2521 14.4571 506 0.6250 0.6014 0.6250 0.7906
0.2521 14.5143 508 0.6477 0.6218 0.6477 0.8048
0.2521 14.5714 510 0.6540 0.6109 0.6540 0.8087
0.2521 14.6286 512 0.6325 0.6109 0.6325 0.7953
0.2521 14.6857 514 0.6086 0.6198 0.6086 0.7801
0.2521 14.7429 516 0.5793 0.6339 0.5793 0.7611
0.2521 14.8 518 0.5804 0.6185 0.5804 0.7618
0.2521 14.8571 520 0.5926 0.6278 0.5926 0.7698
0.2521 14.9143 522 0.6357 0.6035 0.6357 0.7973
0.2521 14.9714 524 0.7009 0.5549 0.7009 0.8372
0.2521 15.0286 526 0.6995 0.5342 0.6995 0.8363
0.2521 15.0857 528 0.6935 0.5267 0.6935 0.8327
0.2521 15.1429 530 0.7112 0.5267 0.7112 0.8433

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k7_task5_organization

Finetuned
(4019)
this model