ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k2_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7585
  • Qwk: 0.5379
  • Mse: 0.7585
  • Rmse: 0.8709

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 4.1292 0.0176 4.1292 2.0320
No log 0.3077 4 2.3388 -0.0040 2.3388 1.5293
No log 0.4615 6 2.0548 0.0159 2.0548 1.4334
No log 0.6154 8 1.3593 0.0 1.3593 1.1659
No log 0.7692 10 1.3094 0.0085 1.3094 1.1443
No log 0.9231 12 1.1318 0.2539 1.1318 1.0639
No log 1.0769 14 1.1443 0.1576 1.1443 1.0697
No log 1.2308 16 1.2703 0.0256 1.2703 1.1271
No log 1.3846 18 1.1505 0.1537 1.1505 1.0726
No log 1.5385 20 1.1270 0.1537 1.1270 1.0616
No log 1.6923 22 1.4146 0.0861 1.4146 1.1894
No log 1.8462 24 1.5335 0.1309 1.5335 1.2383
No log 2.0 26 1.3449 0.1500 1.3449 1.1597
No log 2.1538 28 0.9575 0.1779 0.9575 0.9785
No log 2.3077 30 0.8852 0.3488 0.8852 0.9408
No log 2.4615 32 1.0567 0.2662 1.0567 1.0280
No log 2.6154 34 1.1212 0.2260 1.1212 1.0589
No log 2.7692 36 1.2096 0.2598 1.2096 1.0998
No log 2.9231 38 0.9409 0.3753 0.9409 0.9700
No log 3.0769 40 0.8424 0.4292 0.8424 0.9178
No log 3.2308 42 0.8194 0.4057 0.8194 0.9052
No log 3.3846 44 0.8687 0.4656 0.8687 0.9320
No log 3.5385 46 0.9196 0.4548 0.9196 0.9589
No log 3.6923 48 0.8449 0.4300 0.8449 0.9192
No log 3.8462 50 0.8189 0.5081 0.8189 0.9049
No log 4.0 52 0.7646 0.5275 0.7646 0.8744
No log 4.1538 54 0.7894 0.5287 0.7894 0.8885
No log 4.3077 56 1.1151 0.4576 1.1151 1.0560
No log 4.4615 58 1.2347 0.4410 1.2347 1.1112
No log 4.6154 60 0.8666 0.4984 0.8666 0.9309
No log 4.7692 62 0.8340 0.5182 0.8340 0.9132
No log 4.9231 64 1.0652 0.3942 1.0652 1.0321
No log 5.0769 66 1.1028 0.4191 1.1028 1.0502
No log 5.2308 68 0.9276 0.4616 0.9276 0.9631
No log 5.3846 70 1.1863 0.4398 1.1863 1.0892
No log 5.5385 72 1.2343 0.4351 1.2343 1.1110
No log 5.6923 74 1.4606 0.2939 1.4606 1.2086
No log 5.8462 76 1.3333 0.3529 1.3333 1.1547
No log 6.0 78 1.3359 0.3529 1.3359 1.1558
No log 6.1538 80 1.1767 0.3976 1.1767 1.0848
No log 6.3077 82 0.8492 0.5462 0.8492 0.9215
No log 6.4615 84 0.7134 0.5050 0.7134 0.8447
No log 6.6154 86 0.6935 0.5381 0.6935 0.8328
No log 6.7692 88 0.7186 0.5542 0.7186 0.8477
No log 6.9231 90 0.8803 0.5039 0.8803 0.9382
No log 7.0769 92 0.8352 0.4700 0.8352 0.9139
No log 7.2308 94 1.0424 0.4222 1.0424 1.0210
No log 7.3846 96 1.5415 0.3086 1.5415 1.2416
No log 7.5385 98 1.2893 0.3539 1.2893 1.1355
No log 7.6923 100 0.7130 0.5573 0.7130 0.8444
No log 7.8462 102 0.6472 0.5575 0.6472 0.8045
No log 8.0 104 0.6474 0.5934 0.6474 0.8046
No log 8.1538 106 0.7883 0.5245 0.7883 0.8879
No log 8.3077 108 1.1932 0.4041 1.1932 1.0924
No log 8.4615 110 1.3829 0.3086 1.3829 1.1760
No log 8.6154 112 0.9662 0.5186 0.9662 0.9829
No log 8.7692 114 0.6533 0.5854 0.6533 0.8083
No log 8.9231 116 0.6855 0.4850 0.6855 0.8279
No log 9.0769 118 0.6871 0.4865 0.6871 0.8289
No log 9.2308 120 0.7316 0.5327 0.7316 0.8553
No log 9.3846 122 0.9132 0.5384 0.9132 0.9556
No log 9.5385 124 0.8610 0.5701 0.8610 0.9279
No log 9.6923 126 0.7125 0.5554 0.7125 0.8441
No log 9.8462 128 0.6281 0.5759 0.6281 0.7925
No log 10.0 130 0.6325 0.6089 0.6325 0.7953
No log 10.1538 132 0.8391 0.6239 0.8391 0.9160
No log 10.3077 134 0.9793 0.4718 0.9793 0.9896
No log 10.4615 136 0.8062 0.6260 0.8062 0.8979
No log 10.6154 138 0.6517 0.5961 0.6517 0.8073
No log 10.7692 140 0.6455 0.5820 0.6455 0.8034
No log 10.9231 142 0.6704 0.5626 0.6704 0.8187
No log 11.0769 144 0.9116 0.4537 0.9116 0.9548
No log 11.2308 146 0.8992 0.4754 0.8992 0.9482
No log 11.3846 148 0.6555 0.5843 0.6555 0.8096
No log 11.5385 150 0.6375 0.5660 0.6375 0.7985
No log 11.6923 152 0.6319 0.5660 0.6319 0.7949
No log 11.8462 154 0.6284 0.5843 0.6284 0.7927
No log 12.0 156 0.6223 0.5703 0.6223 0.7888
No log 12.1538 158 0.6362 0.5854 0.6362 0.7976
No log 12.3077 160 0.6891 0.6322 0.6891 0.8301
No log 12.4615 162 0.7469 0.6709 0.7469 0.8642
No log 12.6154 164 0.7083 0.5518 0.7083 0.8416
No log 12.7692 166 0.6901 0.5391 0.6901 0.8307
No log 12.9231 168 0.6681 0.5158 0.6681 0.8174
No log 13.0769 170 0.6740 0.5275 0.6740 0.8210
No log 13.2308 172 0.6606 0.5050 0.6606 0.8128
No log 13.3846 174 0.6993 0.5748 0.6993 0.8362
No log 13.5385 176 0.7030 0.5890 0.7030 0.8385
No log 13.6923 178 0.6718 0.5506 0.6718 0.8196
No log 13.8462 180 0.6472 0.6080 0.6472 0.8045
No log 14.0 182 0.6309 0.6139 0.6309 0.7943
No log 14.1538 184 0.6420 0.6315 0.6420 0.8012
No log 14.3077 186 0.7095 0.6410 0.7095 0.8423
No log 14.4615 188 0.8671 0.5365 0.8671 0.9312
No log 14.6154 190 1.5216 0.3532 1.5216 1.2335
No log 14.7692 192 1.8017 0.2519 1.8017 1.3423
No log 14.9231 194 1.5379 0.2751 1.5379 1.2401
No log 15.0769 196 1.1786 0.4538 1.1786 1.0856
No log 15.2308 198 0.9705 0.5182 0.9705 0.9851
No log 15.3846 200 0.7638 0.4820 0.7638 0.8740
No log 15.5385 202 0.7550 0.4817 0.7550 0.8689
No log 15.6923 204 0.9159 0.5182 0.9159 0.9570
No log 15.8462 206 0.9158 0.5182 0.9158 0.9570
No log 16.0 208 0.7597 0.5148 0.7597 0.8716
No log 16.1538 210 0.6765 0.5759 0.6765 0.8225
No log 16.3077 212 0.6753 0.5080 0.6753 0.8218
No log 16.4615 214 0.6731 0.4955 0.6731 0.8204
No log 16.6154 216 0.6704 0.5274 0.6704 0.8188
No log 16.7692 218 0.6964 0.5463 0.6964 0.8345
No log 16.9231 220 0.7118 0.5442 0.7118 0.8437
No log 17.0769 222 0.6763 0.6052 0.6763 0.8224
No log 17.2308 224 0.6813 0.6052 0.6813 0.8254
No log 17.3846 226 0.7557 0.5749 0.7557 0.8693
No log 17.5385 228 0.9123 0.5184 0.9123 0.9551
No log 17.6923 230 0.9524 0.5184 0.9524 0.9759
No log 17.8462 232 0.8085 0.5134 0.8085 0.8992
No log 18.0 234 0.7235 0.4498 0.7235 0.8506
No log 18.1538 236 0.7109 0.4261 0.7109 0.8431
No log 18.3077 238 0.7379 0.5107 0.7379 0.8590
No log 18.4615 240 0.8269 0.4686 0.8269 0.9093
No log 18.6154 242 0.7918 0.4792 0.7918 0.8898
No log 18.7692 244 0.7098 0.5299 0.7098 0.8425
No log 18.9231 246 0.7031 0.5412 0.7031 0.8385
No log 19.0769 248 0.6846 0.5209 0.6846 0.8274
No log 19.2308 250 0.6759 0.5446 0.6759 0.8222
No log 19.3846 252 0.6773 0.5566 0.6773 0.8230
No log 19.5385 254 0.6761 0.5442 0.6761 0.8222
No log 19.6923 256 0.6860 0.5793 0.6860 0.8282
No log 19.8462 258 0.6186 0.5871 0.6186 0.7865
No log 20.0 260 0.6229 0.5969 0.6229 0.7893
No log 20.1538 262 0.6158 0.5884 0.6158 0.7847
No log 20.3077 264 0.6855 0.6120 0.6855 0.8280
No log 20.4615 266 0.7841 0.5781 0.7841 0.8855
No log 20.6154 268 0.7160 0.5793 0.7160 0.8462
No log 20.7692 270 0.6118 0.5945 0.6118 0.7822
No log 20.9231 272 0.6129 0.5614 0.6129 0.7829
No log 21.0769 274 0.6242 0.5614 0.6242 0.7901
No log 21.2308 276 0.6635 0.5962 0.6635 0.8145
No log 21.3846 278 0.7633 0.4461 0.7633 0.8737
No log 21.5385 280 0.7239 0.5506 0.7239 0.8508
No log 21.6923 282 0.6163 0.6566 0.6163 0.7850
No log 21.8462 284 0.6215 0.5654 0.6215 0.7883
No log 22.0 286 0.7064 0.5675 0.7064 0.8405
No log 22.1538 288 0.6910 0.5475 0.6910 0.8313
No log 22.3077 290 0.6258 0.5763 0.6258 0.7911
No log 22.4615 292 0.6428 0.5674 0.6428 0.8017
No log 22.6154 294 0.7602 0.6018 0.7602 0.8719
No log 22.7692 296 0.7525 0.6018 0.7525 0.8675
No log 22.9231 298 0.6509 0.6099 0.6509 0.8068
No log 23.0769 300 0.6136 0.5714 0.6136 0.7834
No log 23.2308 302 0.6187 0.5830 0.6187 0.7866
No log 23.3846 304 0.6196 0.5466 0.6196 0.7872
No log 23.5385 306 0.6391 0.6259 0.6391 0.7994
No log 23.6923 308 0.6785 0.5487 0.6785 0.8237
No log 23.8462 310 0.6613 0.5602 0.6613 0.8132
No log 24.0 312 0.6226 0.5125 0.6226 0.7890
No log 24.1538 314 0.6514 0.5614 0.6514 0.8071
No log 24.3077 316 0.6727 0.5614 0.6727 0.8202
No log 24.4615 318 0.6324 0.5614 0.6324 0.7952
No log 24.6154 320 0.6023 0.5610 0.6023 0.7761
No log 24.7692 322 0.6184 0.6122 0.6184 0.7864
No log 24.9231 324 0.6101 0.6122 0.6101 0.7811
No log 25.0769 326 0.5905 0.6014 0.5905 0.7684
No log 25.2308 328 0.5870 0.5580 0.5870 0.7661
No log 25.3846 330 0.6021 0.5841 0.6021 0.7760
No log 25.5385 332 0.6035 0.6032 0.6035 0.7768
No log 25.6923 334 0.5908 0.5569 0.5908 0.7686
No log 25.8462 336 0.6563 0.6099 0.6563 0.8101
No log 26.0 338 0.6870 0.5978 0.6870 0.8288
No log 26.1538 340 0.6558 0.5844 0.6558 0.8098
No log 26.3077 342 0.6276 0.5690 0.6276 0.7922
No log 26.4615 344 0.6481 0.5933 0.6481 0.8051
No log 26.6154 346 0.6581 0.5933 0.6581 0.8112
No log 26.7692 348 0.6427 0.5568 0.6427 0.8017
No log 26.9231 350 0.6395 0.6278 0.6395 0.7997
No log 27.0769 352 0.6204 0.5495 0.6204 0.7877
No log 27.2308 354 0.6288 0.5774 0.6288 0.7930
No log 27.3846 356 0.6341 0.5774 0.6341 0.7963
No log 27.5385 358 0.6444 0.5471 0.6444 0.8028
No log 27.6923 360 0.6610 0.4819 0.6610 0.8130
No log 27.8462 362 0.6625 0.4831 0.6625 0.8139
No log 28.0 364 0.6615 0.5192 0.6615 0.8133
No log 28.1538 366 0.6557 0.5580 0.6557 0.8098
No log 28.3077 368 0.6544 0.5763 0.6544 0.8090
No log 28.4615 370 0.6583 0.5657 0.6583 0.8114
No log 28.6154 372 0.6550 0.5450 0.6550 0.8093
No log 28.7692 374 0.6605 0.5024 0.6605 0.8127
No log 28.9231 376 0.6659 0.4960 0.6659 0.8160
No log 29.0769 378 0.6878 0.5368 0.6878 0.8293
No log 29.2308 380 0.7409 0.5381 0.7409 0.8607
No log 29.3846 382 0.7811 0.5471 0.7811 0.8838
No log 29.5385 384 0.7265 0.5536 0.7265 0.8524
No log 29.6923 386 0.6833 0.5580 0.6833 0.8266
No log 29.8462 388 0.6859 0.5224 0.6859 0.8282
No log 30.0 390 0.6829 0.5657 0.6829 0.8264
No log 30.1538 392 0.6643 0.5261 0.6643 0.8151
No log 30.3077 394 0.6819 0.5202 0.6819 0.8258
No log 30.4615 396 0.6906 0.5794 0.6906 0.8310
No log 30.6154 398 0.6606 0.5794 0.6606 0.8128
No log 30.7692 400 0.6245 0.5459 0.6245 0.7903
No log 30.9231 402 0.6144 0.5877 0.6144 0.7838
No log 31.0769 404 0.6164 0.5964 0.6164 0.7851
No log 31.2308 406 0.6058 0.5662 0.6058 0.7783
No log 31.3846 408 0.6359 0.6461 0.6359 0.7974
No log 31.5385 410 0.6747 0.5955 0.6747 0.8214
No log 31.6923 412 0.7503 0.5483 0.7503 0.8662
No log 31.8462 414 0.7140 0.5805 0.7140 0.8450
No log 32.0 416 0.6415 0.6312 0.6415 0.8010
No log 32.1538 418 0.6306 0.6383 0.6306 0.7941
No log 32.3077 420 0.6252 0.5934 0.6252 0.7907
No log 32.4615 422 0.6411 0.6419 0.6411 0.8007
No log 32.6154 424 0.6525 0.5982 0.6525 0.8078
No log 32.7692 426 0.6383 0.6419 0.6383 0.7989
No log 32.9231 428 0.6145 0.6078 0.6145 0.7839
No log 33.0769 430 0.6104 0.6078 0.6104 0.7813
No log 33.2308 432 0.6273 0.6419 0.6273 0.7920
No log 33.3846 434 0.6621 0.5366 0.6621 0.8137
No log 33.5385 436 0.6404 0.6119 0.6404 0.8002
No log 33.6923 438 0.6057 0.6078 0.6057 0.7783
No log 33.8462 440 0.6060 0.6049 0.6060 0.7784
No log 34.0 442 0.6319 0.6119 0.6319 0.7949
No log 34.1538 444 0.6795 0.5924 0.6795 0.8243
No log 34.3077 446 0.6795 0.5471 0.6795 0.8243
No log 34.4615 448 0.6279 0.6119 0.6279 0.7924
No log 34.6154 450 0.5950 0.6237 0.5950 0.7714
No log 34.7692 452 0.5899 0.6237 0.5899 0.7680
No log 34.9231 454 0.6303 0.6099 0.6303 0.7939
No log 35.0769 456 0.6729 0.6226 0.6729 0.8203
No log 35.2308 458 0.6678 0.5887 0.6678 0.8172
No log 35.3846 460 0.6175 0.5854 0.6175 0.7858
No log 35.5385 462 0.6048 0.5854 0.6048 0.7777
No log 35.6923 464 0.6043 0.5854 0.6043 0.7774
No log 35.8462 466 0.6224 0.5748 0.6224 0.7889
No log 36.0 468 0.6296 0.5770 0.6296 0.7935
No log 36.1538 470 0.6110 0.6185 0.6110 0.7817
No log 36.3077 472 0.6109 0.6185 0.6109 0.7816
No log 36.4615 474 0.6285 0.5879 0.6285 0.7928
No log 36.6154 476 0.6449 0.5438 0.6449 0.8030
No log 36.7692 478 0.6202 0.5854 0.6202 0.7875
No log 36.9231 480 0.6058 0.6049 0.6058 0.7783
No log 37.0769 482 0.5992 0.6124 0.5992 0.7741
No log 37.2308 484 0.6019 0.5887 0.6019 0.7758
No log 37.3846 486 0.6104 0.5923 0.6104 0.7813
No log 37.5385 488 0.6291 0.5548 0.6291 0.7932
No log 37.6923 490 0.6503 0.5696 0.6503 0.8064
No log 37.8462 492 0.6348 0.6039 0.6348 0.7967
No log 38.0 494 0.6158 0.5725 0.6158 0.7847
No log 38.1538 496 0.6150 0.5928 0.6150 0.7842
No log 38.3077 498 0.6146 0.5928 0.6146 0.7839
0.2579 38.4615 500 0.6147 0.6046 0.6147 0.7840
0.2579 38.6154 502 0.6225 0.5066 0.6225 0.7890
0.2579 38.7692 504 0.6598 0.5940 0.6598 0.8123
0.2579 38.9231 506 0.6977 0.5940 0.6977 0.8353
0.2579 39.0769 508 0.7571 0.5381 0.7571 0.8701
0.2579 39.2308 510 0.7585 0.5379 0.7585 0.8709

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k2_task5_organization

Finetuned
(4019)
this model