ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k14_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6389
  • Qwk: 0.6120
  • Mse: 0.6389
  • Rmse: 0.7993

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0286 2 3.9489 -0.0307 3.9489 1.9872
No log 0.0571 4 2.1772 0.0985 2.1772 1.4755
No log 0.0857 6 2.2159 -0.0086 2.2159 1.4886
No log 0.1143 8 1.2071 0.2023 1.2071 1.0987
No log 0.1429 10 1.1051 0.0604 1.1051 1.0513
No log 0.1714 12 1.0992 0.0888 1.0992 1.0484
No log 0.2 14 1.0433 0.1799 1.0433 1.0214
No log 0.2286 16 1.0198 0.1767 1.0198 1.0099
No log 0.2571 18 1.0643 0.2880 1.0643 1.0317
No log 0.2857 20 1.0884 0.2100 1.0884 1.0433
No log 0.3143 22 1.0462 0.2689 1.0462 1.0228
No log 0.3429 24 1.0269 0.3646 1.0269 1.0133
No log 0.3714 26 0.9022 0.3915 0.9022 0.9498
No log 0.4 28 0.9047 0.4204 0.9047 0.9512
No log 0.4286 30 1.0539 0.4203 1.0539 1.0266
No log 0.4571 32 1.2357 0.3216 1.2357 1.1116
No log 0.4857 34 1.0594 0.3951 1.0594 1.0293
No log 0.5143 36 0.8061 0.4329 0.8061 0.8978
No log 0.5429 38 0.8269 0.4455 0.8269 0.9093
No log 0.5714 40 0.9144 0.4852 0.9144 0.9562
No log 0.6 42 1.4250 0.3275 1.4250 1.1937
No log 0.6286 44 1.6136 0.3134 1.6136 1.2703
No log 0.6571 46 1.4686 0.3058 1.4686 1.2119
No log 0.6857 48 0.8995 0.5648 0.8995 0.9484
No log 0.7143 50 0.6700 0.5340 0.6700 0.8186
No log 0.7429 52 0.6714 0.4888 0.6714 0.8194
No log 0.7714 54 0.6519 0.6417 0.6519 0.8074
No log 0.8 56 0.7017 0.5717 0.7017 0.8376
No log 0.8286 58 0.6643 0.5542 0.6643 0.8151
No log 0.8571 60 0.6490 0.6196 0.6490 0.8056
No log 0.8857 62 0.6478 0.5996 0.6478 0.8048
No log 0.9143 64 0.6765 0.5855 0.6765 0.8225
No log 0.9429 66 0.6508 0.6399 0.6508 0.8067
No log 0.9714 68 0.7201 0.6524 0.7201 0.8486
No log 1.0 70 0.6879 0.6610 0.6879 0.8294
No log 1.0286 72 0.6219 0.6636 0.6219 0.7886
No log 1.0571 74 0.6577 0.6936 0.6577 0.8110
No log 1.0857 76 0.6130 0.7063 0.6130 0.7829
No log 1.1143 78 0.5769 0.6703 0.5769 0.7595
No log 1.1429 80 0.6644 0.6420 0.6644 0.8151
No log 1.1714 82 0.6364 0.6316 0.6364 0.7978
No log 1.2 84 0.6046 0.6290 0.6046 0.7776
No log 1.2286 86 0.5750 0.6605 0.5750 0.7583
No log 1.2571 88 0.5779 0.6570 0.5779 0.7602
No log 1.2857 90 0.5869 0.6711 0.5869 0.7661
No log 1.3143 92 0.6168 0.6730 0.6168 0.7854
No log 1.3429 94 0.7247 0.6570 0.7247 0.8513
No log 1.3714 96 0.6768 0.6434 0.6768 0.8227
No log 1.4 98 0.6008 0.7104 0.6008 0.7751
No log 1.4286 100 0.5624 0.6589 0.5624 0.7500
No log 1.4571 102 0.5569 0.6689 0.5569 0.7463
No log 1.4857 104 0.5600 0.6589 0.5600 0.7484
No log 1.5143 106 0.5894 0.6998 0.5894 0.7677
No log 1.5429 108 0.7066 0.6439 0.7066 0.8406
No log 1.5714 110 0.7078 0.6696 0.7078 0.8413
No log 1.6 112 0.6057 0.6751 0.6057 0.7782
No log 1.6286 114 0.5749 0.6882 0.5749 0.7582
No log 1.6571 116 0.5873 0.6882 0.5873 0.7663
No log 1.6857 118 0.5835 0.7239 0.5835 0.7639
No log 1.7143 120 0.6080 0.6693 0.6080 0.7797
No log 1.7429 122 0.6120 0.6828 0.6120 0.7823
No log 1.7714 124 0.6464 0.6903 0.6464 0.8040
No log 1.8 126 0.8059 0.6330 0.8059 0.8977
No log 1.8286 128 1.0857 0.4886 1.0857 1.0420
No log 1.8571 130 0.9605 0.6234 0.9605 0.9801
No log 1.8857 132 0.6941 0.6706 0.6941 0.8332
No log 1.9143 134 0.7027 0.5962 0.7027 0.8383
No log 1.9429 136 0.7684 0.5799 0.7684 0.8766
No log 1.9714 138 0.7115 0.6160 0.7115 0.8435
No log 2.0 140 0.6397 0.6259 0.6397 0.7998
No log 2.0286 142 0.7094 0.5339 0.7094 0.8423
No log 2.0571 144 0.7122 0.5430 0.7122 0.8439
No log 2.0857 146 0.6664 0.6039 0.6664 0.8163
No log 2.1143 148 0.8280 0.5422 0.8280 0.9099
No log 2.1429 150 0.9145 0.4894 0.9145 0.9563
No log 2.1714 152 0.8427 0.5106 0.8427 0.9180
No log 2.2 154 0.6977 0.6198 0.6977 0.8353
No log 2.2286 156 0.6506 0.6584 0.6506 0.8066
No log 2.2571 158 0.6165 0.6798 0.6165 0.7852
No log 2.2857 160 0.6451 0.6537 0.6451 0.8032
No log 2.3143 162 0.7602 0.5838 0.7602 0.8719
No log 2.3429 164 0.8142 0.5914 0.8142 0.9023
No log 2.3714 166 0.6528 0.6350 0.6528 0.8080
No log 2.4 168 0.5388 0.6990 0.5388 0.7341
No log 2.4286 170 0.6021 0.6606 0.6021 0.7760
No log 2.4571 172 0.6011 0.7033 0.6011 0.7753
No log 2.4857 174 0.5497 0.6999 0.5497 0.7414
No log 2.5143 176 0.5708 0.7460 0.5708 0.7555
No log 2.5429 178 0.5940 0.6708 0.5940 0.7707
No log 2.5714 180 0.5943 0.6187 0.5943 0.7709
No log 2.6 182 0.6209 0.6325 0.6209 0.7880
No log 2.6286 184 0.6116 0.6187 0.6116 0.7820
No log 2.6571 186 0.6340 0.6986 0.6340 0.7962
No log 2.6857 188 0.6595 0.7195 0.6595 0.8121
No log 2.7143 190 0.6729 0.6978 0.6729 0.8203
No log 2.7429 192 0.6297 0.7073 0.6297 0.7935
No log 2.7714 194 0.6209 0.7189 0.6209 0.7880
No log 2.8 196 0.6087 0.7546 0.6087 0.7802
No log 2.8286 198 0.5666 0.7376 0.5666 0.7527
No log 2.8571 200 0.5643 0.7096 0.5643 0.7512
No log 2.8857 202 0.6194 0.6974 0.6194 0.7870
No log 2.9143 204 0.6417 0.6818 0.6417 0.8011
No log 2.9429 206 0.5761 0.6952 0.5761 0.7590
No log 2.9714 208 0.5562 0.6709 0.5562 0.7458
No log 3.0 210 0.6191 0.7214 0.6191 0.7868
No log 3.0286 212 0.6105 0.7013 0.6105 0.7814
No log 3.0571 214 0.5587 0.7034 0.5587 0.7475
No log 3.0857 216 0.6202 0.6835 0.6202 0.7876
No log 3.1143 218 0.6793 0.6067 0.6793 0.8242
No log 3.1429 220 0.6241 0.6938 0.6241 0.7900
No log 3.1714 222 0.6145 0.7027 0.6145 0.7839
No log 3.2 224 0.6717 0.6054 0.6717 0.8196
No log 3.2286 226 0.6788 0.5356 0.6788 0.8239
No log 3.2571 228 0.6422 0.5150 0.6422 0.8014
No log 3.2857 230 0.6218 0.5708 0.6218 0.7885
No log 3.3143 232 0.5797 0.6055 0.5797 0.7614
No log 3.3429 234 0.5603 0.7077 0.5603 0.7486
No log 3.3714 236 0.5707 0.7077 0.5707 0.7555
No log 3.4 238 0.5710 0.6976 0.5710 0.7556
No log 3.4286 240 0.5547 0.6966 0.5547 0.7448
No log 3.4571 242 0.5507 0.7071 0.5507 0.7421
No log 3.4857 244 0.5553 0.7071 0.5553 0.7452
No log 3.5143 246 0.5664 0.6966 0.5664 0.7526
No log 3.5429 248 0.5974 0.6916 0.5974 0.7729
No log 3.5714 250 0.6239 0.6444 0.6239 0.7899
No log 3.6 252 0.6047 0.6916 0.6047 0.7776
No log 3.6286 254 0.6005 0.6882 0.6005 0.7749
No log 3.6571 256 0.6195 0.6812 0.6195 0.7871
No log 3.6857 258 0.6094 0.6812 0.6094 0.7807
No log 3.7143 260 0.6122 0.6586 0.6122 0.7824
No log 3.7429 262 0.7153 0.6459 0.7153 0.8458
No log 3.7714 264 0.7646 0.5990 0.7646 0.8744
No log 3.8 266 0.6456 0.6649 0.6456 0.8035
No log 3.8286 268 0.5749 0.6947 0.5749 0.7582
No log 3.8571 270 0.6057 0.6507 0.6057 0.7782
No log 3.8857 272 0.5910 0.6822 0.5910 0.7687
No log 3.9143 274 0.5813 0.6699 0.5813 0.7624
No log 3.9429 276 0.5894 0.6846 0.5894 0.7677
No log 3.9714 278 0.6044 0.6843 0.6044 0.7774
No log 4.0 280 0.6219 0.6736 0.6219 0.7886
No log 4.0286 282 0.6356 0.6850 0.6356 0.7972
No log 4.0571 284 0.6282 0.6820 0.6282 0.7926
No log 4.0857 286 0.6765 0.5940 0.6765 0.8225
No log 4.1143 288 0.6910 0.5256 0.6910 0.8313
No log 4.1429 290 0.6789 0.5244 0.6789 0.8239
No log 4.1714 292 0.6139 0.6007 0.6139 0.7835
No log 4.2 294 0.5917 0.6890 0.5917 0.7692
No log 4.2286 296 0.6010 0.6357 0.6010 0.7753
No log 4.2571 298 0.5726 0.6935 0.5726 0.7567
No log 4.2857 300 0.5580 0.6724 0.5580 0.7470
No log 4.3143 302 0.5631 0.6451 0.5631 0.7504
No log 4.3429 304 0.5741 0.6320 0.5741 0.7577
No log 4.3714 306 0.5773 0.6442 0.5773 0.7598
No log 4.4 308 0.6043 0.6035 0.6043 0.7773
No log 4.4286 310 0.5971 0.6330 0.5971 0.7727
No log 4.4571 312 0.5759 0.7165 0.5759 0.7589
No log 4.4857 314 0.5796 0.7124 0.5796 0.7613
No log 4.5143 316 0.5873 0.6966 0.5873 0.7663
No log 4.5429 318 0.6378 0.6263 0.6378 0.7986
No log 4.5714 320 0.7025 0.5930 0.7025 0.8382
No log 4.6 322 0.6932 0.5930 0.6932 0.8326
No log 4.6286 324 0.6365 0.6083 0.6365 0.7978
No log 4.6571 326 0.5887 0.6838 0.5887 0.7673
No log 4.6857 328 0.6088 0.6184 0.6088 0.7803
No log 4.7143 330 0.5996 0.6108 0.5996 0.7743
No log 4.7429 332 0.5968 0.5747 0.5968 0.7725
No log 4.7714 334 0.6429 0.5654 0.6429 0.8018
No log 4.8 336 0.6574 0.6025 0.6574 0.8108
No log 4.8286 338 0.5998 0.6247 0.5998 0.7745
No log 4.8571 340 0.5711 0.6439 0.5711 0.7557
No log 4.8857 342 0.5822 0.6605 0.5822 0.7630
No log 4.9143 344 0.5968 0.6374 0.5968 0.7725
No log 4.9429 346 0.5956 0.6605 0.5956 0.7718
No log 4.9714 348 0.5888 0.6256 0.5888 0.7673
No log 5.0 350 0.5893 0.6094 0.5893 0.7676
No log 5.0286 352 0.5763 0.6690 0.5763 0.7592
No log 5.0571 354 0.5649 0.6874 0.5649 0.7516
No log 5.0857 356 0.5608 0.6909 0.5608 0.7489
No log 5.1143 358 0.5928 0.6536 0.5928 0.7699
No log 5.1429 360 0.6199 0.6657 0.6199 0.7874
No log 5.1714 362 0.6225 0.6560 0.6225 0.7890
No log 5.2 364 0.5760 0.6536 0.5760 0.7590
No log 5.2286 366 0.5692 0.6356 0.5692 0.7544
No log 5.2571 368 0.5859 0.6356 0.5859 0.7654
No log 5.2857 370 0.6380 0.5987 0.6380 0.7987
No log 5.3143 372 0.7020 0.5911 0.7020 0.8378
No log 5.3429 374 0.6792 0.5930 0.6792 0.8241
No log 5.3714 376 0.6328 0.6501 0.6328 0.7955
No log 5.4 378 0.6128 0.6280 0.6128 0.7828
No log 5.4286 380 0.6473 0.6197 0.6473 0.8045
No log 5.4571 382 0.6628 0.6463 0.6628 0.8141
No log 5.4857 384 0.6489 0.6324 0.6489 0.8055
No log 5.5143 386 0.5807 0.6197 0.5807 0.7620
No log 5.5429 388 0.5633 0.5960 0.5633 0.7505
No log 5.5714 390 0.5666 0.6243 0.5666 0.7528
No log 5.6 392 0.5638 0.5960 0.5638 0.7509
No log 5.6286 394 0.5684 0.6133 0.5684 0.7539
No log 5.6571 396 0.6117 0.6063 0.6117 0.7821
No log 5.6857 398 0.6521 0.6035 0.6521 0.8075
No log 5.7143 400 0.6487 0.6035 0.6487 0.8054
No log 5.7429 402 0.6684 0.5595 0.6684 0.8175
No log 5.7714 404 0.6867 0.5647 0.6867 0.8287
No log 5.8 406 0.6460 0.6081 0.6460 0.8038
No log 5.8286 408 0.5620 0.5959 0.5620 0.7497
No log 5.8571 410 0.5372 0.6712 0.5372 0.7329
No log 5.8857 412 0.5240 0.6846 0.5240 0.7239
No log 5.9143 414 0.5196 0.6868 0.5196 0.7209
No log 5.9429 416 0.5523 0.6845 0.5523 0.7432
No log 5.9714 418 0.6059 0.6443 0.6059 0.7784
No log 6.0 420 0.5534 0.6813 0.5534 0.7439
No log 6.0286 422 0.5464 0.6732 0.5464 0.7392
No log 6.0571 424 0.5566 0.6337 0.5566 0.7460
No log 6.0857 426 0.5847 0.6249 0.5847 0.7646
No log 6.1143 428 0.6286 0.6592 0.6286 0.7928
No log 6.1429 430 0.6198 0.6420 0.6198 0.7873
No log 6.1714 432 0.6328 0.6491 0.6328 0.7955
No log 6.2 434 0.6006 0.6122 0.6006 0.7750
No log 6.2286 436 0.5848 0.6259 0.5848 0.7647
No log 6.2571 438 0.5915 0.6404 0.5915 0.7691
No log 6.2857 440 0.5824 0.6748 0.5824 0.7631
No log 6.3143 442 0.5595 0.6617 0.5595 0.7480
No log 6.3429 444 0.5468 0.6421 0.5468 0.7395
No log 6.3714 446 0.5486 0.6389 0.5486 0.7406
No log 6.4 448 0.5761 0.6380 0.5761 0.7590
No log 6.4286 450 0.6334 0.6455 0.6334 0.7959
No log 6.4571 452 0.6806 0.6015 0.6806 0.8250
No log 6.4857 454 0.7157 0.5942 0.7157 0.8460
No log 6.5143 456 0.6887 0.5763 0.6887 0.8299
No log 6.5429 458 0.6300 0.5825 0.6300 0.7937
No log 6.5714 460 0.5739 0.5960 0.5739 0.7576
No log 6.6 462 0.5564 0.6139 0.5564 0.7459
No log 6.6286 464 0.5526 0.6822 0.5526 0.7434
No log 6.6571 466 0.5529 0.6593 0.5529 0.7436
No log 6.6857 468 0.5708 0.5969 0.5708 0.7555
No log 6.7143 470 0.6219 0.6601 0.6219 0.7886
No log 6.7429 472 0.6431 0.6632 0.6431 0.8019
No log 6.7714 474 0.6053 0.6305 0.6053 0.7780
No log 6.8 476 0.5781 0.6424 0.5781 0.7604
No log 6.8286 478 0.5769 0.6229 0.5769 0.7595
No log 6.8571 480 0.5704 0.6649 0.5704 0.7553
No log 6.8857 482 0.5651 0.6509 0.5651 0.7517
No log 6.9143 484 0.5836 0.6597 0.5836 0.7639
No log 6.9429 486 0.5997 0.6421 0.5997 0.7744
No log 6.9714 488 0.6185 0.6446 0.6185 0.7864
No log 7.0 490 0.5990 0.6335 0.5990 0.7740
No log 7.0286 492 0.5808 0.6307 0.5808 0.7621
No log 7.0571 494 0.5570 0.6470 0.5570 0.7463
No log 7.0857 496 0.5483 0.6658 0.5483 0.7405
No log 7.1143 498 0.5485 0.6846 0.5485 0.7406
0.2704 7.1429 500 0.5579 0.6740 0.5579 0.7469
0.2704 7.1714 502 0.5717 0.6664 0.5717 0.7561
0.2704 7.2 504 0.6018 0.6427 0.6018 0.7758
0.2704 7.2286 506 0.5778 0.6352 0.5778 0.7601
0.2704 7.2571 508 0.5739 0.6526 0.5739 0.7575
0.2704 7.2857 510 0.6389 0.6120 0.6389 0.7993

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k14_task5_organization

Finetuned
(4019)
this model