ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run999_AugV5_k1_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7527
  • Qwk: 0.6277
  • Mse: 0.7527
  • Rmse: 0.8676

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.25 2 4.5672 -0.0103 4.5672 2.1371
No log 0.5 4 3.8256 0.0165 3.8256 1.9559
No log 0.75 6 1.9924 0.0879 1.9924 1.4115
No log 1.0 8 1.3135 0.0992 1.3135 1.1461
No log 1.25 10 1.1989 0.2532 1.1989 1.0950
No log 1.5 12 1.1159 0.3663 1.1159 1.0564
No log 1.75 14 1.2979 0.1395 1.2979 1.1392
No log 2.0 16 1.7774 0.2007 1.7774 1.3332
No log 2.25 18 1.8266 0.2007 1.8266 1.3515
No log 2.5 20 1.2240 0.2094 1.2240 1.1064
No log 2.75 22 1.1697 0.2991 1.1697 1.0815
No log 3.0 24 1.3357 0.3487 1.3357 1.1557
No log 3.25 26 1.4775 0.3467 1.4775 1.2155
No log 3.5 28 1.3308 0.3337 1.3308 1.1536
No log 3.75 30 1.1887 0.3662 1.1887 1.0903
No log 4.0 32 1.1607 0.4181 1.1607 1.0774
No log 4.25 34 1.1590 0.4145 1.1590 1.0766
No log 4.5 36 1.3189 0.3624 1.3189 1.1484
No log 4.75 38 1.1902 0.3496 1.1902 1.0910
No log 5.0 40 0.9968 0.5283 0.9968 0.9984
No log 5.25 42 0.9532 0.4628 0.9532 0.9763
No log 5.5 44 0.8992 0.5336 0.8992 0.9482
No log 5.75 46 0.9013 0.5661 0.9013 0.9494
No log 6.0 48 1.0160 0.3812 1.0160 1.0080
No log 6.25 50 1.0665 0.2857 1.0665 1.0327
No log 6.5 52 0.9579 0.5161 0.9579 0.9787
No log 6.75 54 0.9063 0.5540 0.9063 0.9520
No log 7.0 56 0.8729 0.5469 0.8729 0.9343
No log 7.25 58 0.9195 0.5925 0.9195 0.9589
No log 7.5 60 0.9483 0.5981 0.9483 0.9738
No log 7.75 62 0.8631 0.5727 0.8631 0.9290
No log 8.0 64 0.8401 0.5316 0.8401 0.9166
No log 8.25 66 0.8532 0.5621 0.8532 0.9237
No log 8.5 68 0.9241 0.5763 0.9241 0.9613
No log 8.75 70 0.8741 0.6476 0.8741 0.9349
No log 9.0 72 0.9365 0.5653 0.9365 0.9677
No log 9.25 74 1.1255 0.3949 1.1255 1.0609
No log 9.5 76 1.1053 0.4032 1.1053 1.0514
No log 9.75 78 0.9055 0.5872 0.9055 0.9516
No log 10.0 80 0.8272 0.6038 0.8272 0.9095
No log 10.25 82 0.7976 0.5886 0.7976 0.8931
No log 10.5 84 0.8105 0.6010 0.8105 0.9003
No log 10.75 86 0.9892 0.5072 0.9892 0.9946
No log 11.0 88 1.1295 0.3995 1.1295 1.0628
No log 11.25 90 1.1286 0.3995 1.1286 1.0623
No log 11.5 92 0.9918 0.5387 0.9918 0.9959
No log 11.75 94 0.9009 0.5643 0.9009 0.9492
No log 12.0 96 0.8678 0.5634 0.8678 0.9316
No log 12.25 98 0.8491 0.5527 0.8491 0.9215
No log 12.5 100 0.9288 0.5848 0.9288 0.9638
No log 12.75 102 1.0346 0.3845 1.0346 1.0172
No log 13.0 104 1.0993 0.3757 1.0993 1.0485
No log 13.25 106 1.0666 0.3757 1.0666 1.0328
No log 13.5 108 0.9152 0.5458 0.9152 0.9567
No log 13.75 110 0.8455 0.5159 0.8455 0.9195
No log 14.0 112 0.8476 0.5012 0.8476 0.9206
No log 14.25 114 0.8169 0.5769 0.8169 0.9038
No log 14.5 116 0.8390 0.5889 0.8390 0.9160
No log 14.75 118 0.8586 0.5889 0.8586 0.9266
No log 15.0 120 0.9263 0.5855 0.9263 0.9624
No log 15.25 122 0.9487 0.5090 0.9487 0.9740
No log 15.5 124 0.9345 0.5202 0.9345 0.9667
No log 15.75 126 0.9169 0.6106 0.9169 0.9576
No log 16.0 128 0.8941 0.5750 0.8941 0.9456
No log 16.25 130 0.8700 0.5706 0.8700 0.9327
No log 16.5 132 0.9050 0.5911 0.9050 0.9513
No log 16.75 134 0.9342 0.5479 0.9342 0.9665
No log 17.0 136 0.9597 0.4526 0.9597 0.9796
No log 17.25 138 0.9473 0.4449 0.9473 0.9733
No log 17.5 140 0.9224 0.5398 0.9224 0.9604
No log 17.75 142 0.8434 0.5835 0.8434 0.9184
No log 18.0 144 0.7839 0.6044 0.7839 0.8854
No log 18.25 146 0.8347 0.5835 0.8347 0.9136
No log 18.5 148 0.8247 0.6060 0.8247 0.9081
No log 18.75 150 0.7776 0.6328 0.7776 0.8818
No log 19.0 152 0.8599 0.5963 0.8599 0.9273
No log 19.25 154 0.9787 0.4765 0.9787 0.9893
No log 19.5 156 0.9388 0.5530 0.9388 0.9689
No log 19.75 158 0.8001 0.5683 0.8001 0.8945
No log 20.0 160 0.7520 0.5611 0.7520 0.8672
No log 20.25 162 0.7939 0.5519 0.7939 0.8910
No log 20.5 164 0.7527 0.5988 0.7527 0.8676
No log 20.75 166 0.7427 0.5993 0.7427 0.8618
No log 21.0 168 0.8660 0.5778 0.8660 0.9306
No log 21.25 170 0.9971 0.4767 0.9971 0.9986
No log 21.5 172 0.9048 0.5892 0.9048 0.9512
No log 21.75 174 0.7944 0.5968 0.7944 0.8913
No log 22.0 176 0.7774 0.5573 0.7774 0.8817
No log 22.25 178 0.7796 0.5621 0.7796 0.8829
No log 22.5 180 0.8166 0.5752 0.8166 0.9037
No log 22.75 182 0.8159 0.5167 0.8159 0.9033
No log 23.0 184 0.8132 0.5413 0.8132 0.9018
No log 23.25 186 0.8279 0.5548 0.8279 0.9099
No log 23.5 188 0.8567 0.6014 0.8567 0.9256
No log 23.75 190 0.8828 0.6067 0.8828 0.9396
No log 24.0 192 0.9324 0.5687 0.9324 0.9656
No log 24.25 194 0.9485 0.5392 0.9485 0.9739
No log 24.5 196 0.8357 0.5911 0.8357 0.9142
No log 24.75 198 0.7682 0.5413 0.7682 0.8765
No log 25.0 200 0.7771 0.4889 0.7771 0.8815
No log 25.25 202 0.7889 0.4889 0.7889 0.8882
No log 25.5 204 0.7767 0.4889 0.7767 0.8813
No log 25.75 206 0.7653 0.5481 0.7653 0.8748
No log 26.0 208 0.7854 0.5969 0.7854 0.8862
No log 26.25 210 0.9035 0.5737 0.9035 0.9505
No log 26.5 212 0.9183 0.5920 0.9183 0.9583
No log 26.75 214 0.8388 0.6212 0.8388 0.9158
No log 27.0 216 0.7822 0.6139 0.7822 0.8844
No log 27.25 218 0.7722 0.6086 0.7722 0.8787
No log 27.5 220 0.8329 0.6095 0.8329 0.9127
No log 27.75 222 0.9132 0.5228 0.9132 0.9556
No log 28.0 224 0.9593 0.5228 0.9593 0.9795
No log 28.25 226 0.9844 0.5148 0.9844 0.9922
No log 28.5 228 0.9307 0.5383 0.9307 0.9647
No log 28.75 230 0.8670 0.6201 0.8670 0.9311
No log 29.0 232 0.7697 0.5854 0.7697 0.8773
No log 29.25 234 0.7627 0.5012 0.7627 0.8733
No log 29.5 236 0.7568 0.5239 0.7568 0.8699
No log 29.75 238 0.7426 0.5315 0.7426 0.8618
No log 30.0 240 0.7565 0.5581 0.7565 0.8698
No log 30.25 242 0.7949 0.5226 0.7949 0.8916
No log 30.5 244 0.7903 0.5279 0.7903 0.8890
No log 30.75 246 0.7935 0.5581 0.7935 0.8908
No log 31.0 248 0.7920 0.5581 0.7920 0.8899
No log 31.25 250 0.7911 0.5581 0.7911 0.8895
No log 31.5 252 0.8080 0.5226 0.8080 0.8989
No log 31.75 254 0.7824 0.5279 0.7824 0.8845
No log 32.0 256 0.7594 0.5830 0.7594 0.8714
No log 32.25 258 0.7512 0.5773 0.7512 0.8667
No log 32.5 260 0.7516 0.5458 0.7516 0.8670
No log 32.75 262 0.7596 0.5944 0.7596 0.8715
No log 33.0 264 0.7556 0.5773 0.7556 0.8693
No log 33.25 266 0.7606 0.5462 0.7606 0.8721
No log 33.5 268 0.7813 0.5443 0.7813 0.8839
No log 33.75 270 0.7974 0.5633 0.7974 0.8930
No log 34.0 272 0.8128 0.5787 0.8128 0.9016
No log 34.25 274 0.7962 0.5691 0.7962 0.8923
No log 34.5 276 0.8200 0.6026 0.8200 0.9055
No log 34.75 278 0.8420 0.6167 0.8420 0.9176
No log 35.0 280 0.8093 0.6167 0.8093 0.8996
No log 35.25 282 0.7566 0.6032 0.7566 0.8698
No log 35.5 284 0.7621 0.5681 0.7621 0.8730
No log 35.75 286 0.7825 0.5956 0.7825 0.8846
No log 36.0 288 0.7786 0.5816 0.7786 0.8824
No log 36.25 290 0.7601 0.5530 0.7601 0.8718
No log 36.5 292 0.7678 0.5194 0.7678 0.8762
No log 36.75 294 0.8233 0.5926 0.8233 0.9073
No log 37.0 296 0.8563 0.5571 0.8563 0.9254
No log 37.25 298 0.8415 0.5513 0.8415 0.9173
No log 37.5 300 0.7797 0.5563 0.7797 0.8830
No log 37.75 302 0.7475 0.5596 0.7475 0.8646
No log 38.0 304 0.7587 0.5915 0.7587 0.8710
No log 38.25 306 0.7819 0.6277 0.7819 0.8843
No log 38.5 308 0.8275 0.6097 0.8275 0.9097
No log 38.75 310 0.8783 0.6321 0.8783 0.9372
No log 39.0 312 0.8847 0.6340 0.8847 0.9406
No log 39.25 314 0.8524 0.6136 0.8524 0.9232
No log 39.5 316 0.8067 0.5971 0.8067 0.8981
No log 39.75 318 0.7862 0.5658 0.7862 0.8867
No log 40.0 320 0.7645 0.5397 0.7645 0.8744
No log 40.25 322 0.7652 0.5582 0.7652 0.8747
No log 40.5 324 0.7895 0.5693 0.7895 0.8886
No log 40.75 326 0.8312 0.6011 0.8312 0.9117
No log 41.0 328 0.8201 0.6074 0.8201 0.9056
No log 41.25 330 0.8093 0.6074 0.8093 0.8996
No log 41.5 332 0.7577 0.5759 0.7577 0.8705
No log 41.75 334 0.7257 0.6108 0.7257 0.8519
No log 42.0 336 0.7305 0.5148 0.7305 0.8547
No log 42.25 338 0.7625 0.5359 0.7625 0.8732
No log 42.5 340 0.7724 0.5515 0.7724 0.8789
No log 42.75 342 0.7583 0.5481 0.7583 0.8708
No log 43.0 344 0.7621 0.6218 0.7621 0.8730
No log 43.25 346 0.8241 0.5783 0.8241 0.9078
No log 43.5 348 0.8800 0.6305 0.8800 0.9381
No log 43.75 350 0.9163 0.6274 0.9163 0.9572
No log 44.0 352 0.9326 0.6241 0.9326 0.9657
No log 44.25 354 0.8969 0.6434 0.8969 0.9470
No log 44.5 356 0.8530 0.6098 0.8530 0.9236
No log 44.75 358 0.8156 0.6108 0.8156 0.9031
No log 45.0 360 0.7856 0.5573 0.7856 0.8864
No log 45.25 362 0.7766 0.5391 0.7766 0.8813
No log 45.5 364 0.7773 0.5396 0.7773 0.8816
No log 45.75 366 0.7936 0.6078 0.7936 0.8908
No log 46.0 368 0.8269 0.6151 0.8269 0.9094
No log 46.25 370 0.8661 0.6026 0.8661 0.9306
No log 46.5 372 0.8898 0.5739 0.8898 0.9433
No log 46.75 374 0.9149 0.5763 0.9149 0.9565
No log 47.0 376 0.8884 0.5763 0.8884 0.9426
No log 47.25 378 0.8597 0.5816 0.8597 0.9272
No log 47.5 380 0.8038 0.6048 0.8038 0.8966
No log 47.75 382 0.7453 0.5930 0.7453 0.8633
No log 48.0 384 0.7387 0.5958 0.7387 0.8595
No log 48.25 386 0.7387 0.5793 0.7387 0.8595
No log 48.5 388 0.7377 0.6075 0.7377 0.8589
No log 48.75 390 0.7344 0.5828 0.7344 0.8570
No log 49.0 392 0.7425 0.5793 0.7425 0.8617
No log 49.25 394 0.7764 0.5976 0.7764 0.8811
No log 49.5 396 0.8145 0.6157 0.8145 0.9025
No log 49.75 398 0.8599 0.5856 0.8599 0.9273
No log 50.0 400 0.9296 0.5533 0.9296 0.9642
No log 50.25 402 0.9645 0.5681 0.9645 0.9821
No log 50.5 404 0.9089 0.5681 0.9089 0.9534
No log 50.75 406 0.8218 0.6202 0.8218 0.9065
No log 51.0 408 0.7478 0.6172 0.7478 0.8648
No log 51.25 410 0.7114 0.6151 0.7114 0.8435
No log 51.5 412 0.6909 0.6487 0.6909 0.8312
No log 51.75 414 0.6946 0.6244 0.6946 0.8334
No log 52.0 416 0.7115 0.6044 0.7115 0.8435
No log 52.25 418 0.7130 0.6228 0.7130 0.8444
No log 52.5 420 0.7106 0.6404 0.7106 0.8430
No log 52.75 422 0.7094 0.6089 0.7094 0.8423
No log 53.0 424 0.7039 0.6629 0.7039 0.8390
No log 53.25 426 0.6985 0.6237 0.6985 0.8358
No log 53.5 428 0.7047 0.5988 0.7047 0.8395
No log 53.75 430 0.7138 0.6328 0.7138 0.8449
No log 54.0 432 0.7195 0.5815 0.7195 0.8482
No log 54.25 434 0.7271 0.6097 0.7271 0.8527
No log 54.5 436 0.7365 0.6647 0.7365 0.8582
No log 54.75 438 0.7400 0.6669 0.7400 0.8602
No log 55.0 440 0.7317 0.6343 0.7317 0.8554
No log 55.25 442 0.7222 0.6343 0.7222 0.8498
No log 55.5 444 0.7139 0.6218 0.7139 0.8450
No log 55.75 446 0.7090 0.6468 0.7090 0.8420
No log 56.0 448 0.7157 0.6393 0.7157 0.8460
No log 56.25 450 0.7200 0.6468 0.7200 0.8486
No log 56.5 452 0.7242 0.6119 0.7242 0.8510
No log 56.75 454 0.7298 0.6119 0.7298 0.8543
No log 57.0 456 0.7402 0.5974 0.7402 0.8604
No log 57.25 458 0.7547 0.6044 0.7547 0.8687
No log 57.5 460 0.7690 0.5931 0.7690 0.8769
No log 57.75 462 0.7846 0.6151 0.7846 0.8858
No log 58.0 464 0.8249 0.6160 0.8249 0.9082
No log 58.25 466 0.8499 0.6146 0.8499 0.9219
No log 58.5 468 0.8405 0.6404 0.8405 0.9168
No log 58.75 470 0.8178 0.6246 0.8178 0.9043
No log 59.0 472 0.7906 0.6225 0.7906 0.8892
No log 59.25 474 0.7889 0.6228 0.7889 0.8882
No log 59.5 476 0.7943 0.6024 0.7943 0.8912
No log 59.75 478 0.7907 0.6024 0.7907 0.8892
No log 60.0 480 0.7812 0.6024 0.7812 0.8838
No log 60.25 482 0.7588 0.5827 0.7588 0.8711
No log 60.5 484 0.7538 0.6343 0.7538 0.8682
No log 60.75 486 0.7566 0.6065 0.7566 0.8699
No log 61.0 488 0.7507 0.6065 0.7507 0.8664
No log 61.25 490 0.7477 0.6205 0.7477 0.8647
No log 61.5 492 0.7367 0.6205 0.7367 0.8583
No log 61.75 494 0.7180 0.6097 0.7180 0.8474
No log 62.0 496 0.7051 0.6257 0.7051 0.8397
No log 62.25 498 0.7055 0.6479 0.7055 0.8399
0.2158 62.5 500 0.7168 0.7044 0.7168 0.8466
0.2158 62.75 502 0.7305 0.6725 0.7305 0.8547
0.2158 63.0 504 0.7416 0.6414 0.7416 0.8611
0.2158 63.25 506 0.7493 0.6414 0.7493 0.8656
0.2158 63.5 508 0.7530 0.6366 0.7530 0.8678
0.2158 63.75 510 0.7527 0.6277 0.7527 0.8676

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run999_AugV5_k1_task2_organization

Finetuned
(4019)
this model