ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k10_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9447
  • Qwk: 0.2615
  • Mse: 0.9447
  • Rmse: 0.9719

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0833 2 2.5562 -0.0449 2.5562 1.5988
No log 0.1667 4 1.2422 0.0726 1.2422 1.1145
No log 0.25 6 0.9587 -0.0970 0.9587 0.9791
No log 0.3333 8 0.8678 0.1010 0.8678 0.9316
No log 0.4167 10 0.7701 0.0688 0.7701 0.8776
No log 0.5 12 0.8958 0.2387 0.8958 0.9465
No log 0.5833 14 0.7668 0.2261 0.7668 0.8757
No log 0.6667 16 0.8024 0.2740 0.8024 0.8957
No log 0.75 18 1.1118 0.0518 1.1118 1.0544
No log 0.8333 20 1.0401 0.2037 1.0401 1.0198
No log 0.9167 22 0.8404 0.1786 0.8404 0.9167
No log 1.0 24 0.8079 0.0937 0.8079 0.8988
No log 1.0833 26 0.8021 0.0 0.8021 0.8956
No log 1.1667 28 0.7711 0.0 0.7711 0.8781
No log 1.25 30 0.7354 0.0 0.7354 0.8576
No log 1.3333 32 0.7161 0.0840 0.7161 0.8462
No log 1.4167 34 0.7131 0.0840 0.7131 0.8444
No log 1.5 36 0.6958 0.1236 0.6958 0.8342
No log 1.5833 38 0.6690 0.3323 0.6690 0.8179
No log 1.6667 40 0.8661 0.3231 0.8661 0.9306
No log 1.75 42 1.0736 0.2510 1.0736 1.0362
No log 1.8333 44 1.1826 -0.0960 1.1826 1.0875
No log 1.9167 46 1.0100 -0.1823 1.0100 1.0050
No log 2.0 48 0.7369 -0.0027 0.7369 0.8585
No log 2.0833 50 0.7957 0.1372 0.7957 0.8920
No log 2.1667 52 0.8417 0.2526 0.8417 0.9175
No log 2.25 54 0.7856 0.2181 0.7856 0.8864
No log 2.3333 56 0.7029 0.0937 0.7029 0.8384
No log 2.4167 58 0.6646 0.0393 0.6646 0.8153
No log 2.5 60 0.7157 0.2817 0.7157 0.8460
No log 2.5833 62 0.6541 0.3789 0.6541 0.8087
No log 2.6667 64 0.5738 0.3416 0.5738 0.7575
No log 2.75 66 0.5736 0.3745 0.5736 0.7574
No log 2.8333 68 0.6010 0.4243 0.6010 0.7753
No log 2.9167 70 0.6401 0.4728 0.6401 0.8001
No log 3.0 72 0.6577 0.4644 0.6577 0.8110
No log 3.0833 74 0.6102 0.4020 0.6102 0.7812
No log 3.1667 76 0.5533 0.4929 0.5533 0.7438
No log 3.25 78 0.5857 0.4259 0.5857 0.7653
No log 3.3333 80 0.5910 0.4259 0.5910 0.7687
No log 3.4167 82 0.5542 0.4161 0.5542 0.7444
No log 3.5 84 0.6259 0.4618 0.6259 0.7911
No log 3.5833 86 0.6814 0.4270 0.6814 0.8255
No log 3.6667 88 0.7034 0.3399 0.7034 0.8387
No log 3.75 90 0.7211 0.3099 0.7211 0.8492
No log 3.8333 92 0.6896 0.2171 0.6896 0.8304
No log 3.9167 94 0.6397 0.2852 0.6397 0.7998
No log 4.0 96 0.6207 0.2783 0.6207 0.7879
No log 4.0833 98 0.7376 0.3712 0.7376 0.8588
No log 4.1667 100 0.8024 0.3782 0.8024 0.8958
No log 4.25 102 0.8846 0.3560 0.8846 0.9405
No log 4.3333 104 0.7927 0.4597 0.7927 0.8903
No log 4.4167 106 0.7750 0.4265 0.7750 0.8803
No log 4.5 108 0.7954 0.4057 0.7954 0.8918
No log 4.5833 110 0.8548 0.4199 0.8548 0.9246
No log 4.6667 112 0.9363 0.3274 0.9363 0.9676
No log 4.75 114 0.9530 0.3274 0.9530 0.9762
No log 4.8333 116 1.0665 0.3007 1.0665 1.0327
No log 4.9167 118 0.9912 0.3174 0.9912 0.9956
No log 5.0 120 0.8297 0.2564 0.8297 0.9109
No log 5.0833 122 0.8274 0.2589 0.8274 0.9096
No log 5.1667 124 1.0500 0.2348 1.0500 1.0247
No log 5.25 126 1.3406 0.2441 1.3406 1.1578
No log 5.3333 128 1.4614 0.2178 1.4614 1.2089
No log 5.4167 130 1.1660 0.2421 1.1660 1.0798
No log 5.5 132 0.9478 0.2706 0.9478 0.9736
No log 5.5833 134 1.1660 0.3211 1.1660 1.0798
No log 5.6667 136 1.6566 0.2552 1.6566 1.2871
No log 5.75 138 1.6510 0.2382 1.6510 1.2849
No log 5.8333 140 1.1293 0.3237 1.1293 1.0627
No log 5.9167 142 0.6346 0.4473 0.6346 0.7966
No log 6.0 144 0.6518 0.3879 0.6518 0.8074
No log 6.0833 146 0.6581 0.3060 0.6581 0.8112
No log 6.1667 148 0.6233 0.5056 0.6233 0.7895
No log 6.25 150 0.7600 0.3869 0.7600 0.8718
No log 6.3333 152 0.9092 0.3029 0.9092 0.9535
No log 6.4167 154 0.8202 0.3251 0.8202 0.9057
No log 6.5 156 0.7005 0.4684 0.7005 0.8370
No log 6.5833 158 0.6666 0.4300 0.6666 0.8165
No log 6.6667 160 0.6864 0.4091 0.6864 0.8285
No log 6.75 162 0.8087 0.3630 0.8087 0.8993
No log 6.8333 164 1.0870 0.2398 1.0870 1.0426
No log 6.9167 166 1.1503 0.2665 1.1503 1.0725
No log 7.0 168 0.9309 0.3019 0.9309 0.9648
No log 7.0833 170 0.8108 0.3409 0.8108 0.9005
No log 7.1667 172 0.7310 0.3341 0.7310 0.8550
No log 7.25 174 0.7451 0.3622 0.7451 0.8632
No log 7.3333 176 0.8579 0.2643 0.8579 0.9263
No log 7.4167 178 0.9580 0.2461 0.9580 0.9788
No log 7.5 180 0.8771 0.2602 0.8771 0.9365
No log 7.5833 182 0.7625 0.3399 0.7625 0.8732
No log 7.6667 184 0.7594 0.3399 0.7594 0.8715
No log 7.75 186 0.8368 0.3043 0.8368 0.9148
No log 7.8333 188 0.8648 0.3194 0.8648 0.9300
No log 7.9167 190 0.9845 0.2297 0.9845 0.9922
No log 8.0 192 0.9908 0.2297 0.9908 0.9954
No log 8.0833 194 0.9315 0.3076 0.9315 0.9652
No log 8.1667 196 0.8757 0.3194 0.8757 0.9358
No log 8.25 198 0.8273 0.2492 0.8273 0.9096
No log 8.3333 200 0.9201 0.3134 0.9201 0.9592
No log 8.4167 202 1.1019 0.2306 1.1019 1.0497
No log 8.5 204 1.2641 0.2020 1.2641 1.1243
No log 8.5833 206 1.1434 0.1815 1.1434 1.0693
No log 8.6667 208 0.8680 0.2905 0.8680 0.9317
No log 8.75 210 0.7248 0.4522 0.7248 0.8513
No log 8.8333 212 0.6911 0.4542 0.6911 0.8313
No log 8.9167 214 0.6927 0.4294 0.6927 0.8323
No log 9.0 216 0.7722 0.4315 0.7722 0.8788
No log 9.0833 218 0.7048 0.3963 0.7048 0.8395
No log 9.1667 220 0.6719 0.3662 0.6719 0.8197
No log 9.25 222 0.6340 0.3622 0.6340 0.7962
No log 9.3333 224 0.6491 0.3545 0.6491 0.8057
No log 9.4167 226 0.6905 0.3127 0.6905 0.8310
No log 9.5 228 0.8063 0.3319 0.8063 0.8979
No log 9.5833 230 0.9515 0.3501 0.9515 0.9754
No log 9.6667 232 1.0571 0.3557 1.0571 1.0282
No log 9.75 234 1.0299 0.3557 1.0299 1.0148
No log 9.8333 236 0.8685 0.2958 0.8685 0.9319
No log 9.9167 238 0.7524 0.3195 0.7524 0.8674
No log 10.0 240 0.7499 0.3261 0.7499 0.8660
No log 10.0833 242 0.7793 0.3195 0.7793 0.8828
No log 10.1667 244 0.8742 0.3606 0.8742 0.9350
No log 10.25 246 0.8508 0.3042 0.8508 0.9224
No log 10.3333 248 0.8179 0.2784 0.8179 0.9044
No log 10.4167 250 0.9100 0.3347 0.9100 0.9539
No log 10.5 252 0.9320 0.3477 0.9320 0.9654
No log 10.5833 254 0.8915 0.4277 0.8915 0.9442
No log 10.6667 256 0.8065 0.3754 0.8065 0.8980
No log 10.75 258 0.7106 0.3095 0.7106 0.8430
No log 10.8333 260 0.7511 0.3195 0.7511 0.8667
No log 10.9167 262 0.8599 0.4255 0.8599 0.9273
No log 11.0 264 0.9451 0.3636 0.9451 0.9722
No log 11.0833 266 0.8938 0.4080 0.8938 0.9454
No log 11.1667 268 0.8620 0.4154 0.8620 0.9284
No log 11.25 270 0.7880 0.4067 0.7880 0.8877
No log 11.3333 272 0.6752 0.2817 0.6752 0.8217
No log 11.4167 274 0.5920 0.3809 0.5920 0.7694
No log 11.5 276 0.5865 0.4001 0.5865 0.7659
No log 11.5833 278 0.6727 0.3127 0.6727 0.8202
No log 11.6667 280 0.9230 0.3719 0.9230 0.9607
No log 11.75 282 1.0803 0.3059 1.0803 1.0394
No log 11.8333 284 1.1370 0.2622 1.1370 1.0663
No log 11.9167 286 0.9791 0.3359 0.9791 0.9895
No log 12.0 288 0.7572 0.4329 0.7572 0.8702
No log 12.0833 290 0.6496 0.3127 0.6496 0.8060
No log 12.1667 292 0.6436 0.3127 0.6436 0.8022
No log 12.25 294 0.6787 0.3127 0.6787 0.8238
No log 12.3333 296 0.8330 0.4230 0.8330 0.9127
No log 12.4167 298 0.9653 0.3359 0.9653 0.9825
No log 12.5 300 0.9905 0.3302 0.9905 0.9953
No log 12.5833 302 0.9474 0.3657 0.9474 0.9734
No log 12.6667 304 0.8886 0.3847 0.8886 0.9427
No log 12.75 306 0.8056 0.4014 0.8056 0.8976
No log 12.8333 308 0.8355 0.3819 0.8355 0.9140
No log 12.9167 310 0.9611 0.3825 0.9611 0.9804
No log 13.0 312 1.0171 0.2898 1.0171 1.0085
No log 13.0833 314 1.0009 0.3082 1.0009 1.0005
No log 13.1667 316 1.0734 0.2876 1.0734 1.0360
No log 13.25 318 1.0299 0.2659 1.0299 1.0148
No log 13.3333 320 0.8969 0.4255 0.8969 0.9470
No log 13.4167 322 0.8775 0.4154 0.8775 0.9368
No log 13.5 324 0.9250 0.2939 0.9250 0.9618
No log 13.5833 326 1.0115 0.2732 1.0115 1.0057
No log 13.6667 328 1.0748 0.2543 1.0748 1.0367
No log 13.75 330 1.0053 0.2977 1.0053 1.0027
No log 13.8333 332 0.8991 0.3731 0.8991 0.9482
No log 13.9167 334 0.8586 0.4275 0.8586 0.9266
No log 14.0 336 0.9507 0.3913 0.9507 0.9751
No log 14.0833 338 1.0505 0.2894 1.0505 1.0250
No log 14.1667 340 1.0796 0.2846 1.0796 1.0390
No log 14.25 342 1.0254 0.3258 1.0254 1.0126
No log 14.3333 344 0.9322 0.3417 0.9322 0.9655
No log 14.4167 346 0.8263 0.2722 0.8263 0.9090
No log 14.5 348 0.8111 0.2440 0.8111 0.9006
No log 14.5833 350 0.8744 0.2995 0.8744 0.9351
No log 14.6667 352 0.9487 0.2439 0.9487 0.9740
No log 14.75 354 0.9808 0.2287 0.9808 0.9904
No log 14.8333 356 1.0231 0.2754 1.0231 1.0115
No log 14.9167 358 1.0308 0.2830 1.0308 1.0153
No log 15.0 360 1.0131 0.2756 1.0131 1.0065
No log 15.0833 362 0.9772 0.2259 0.9772 0.9885
No log 15.1667 364 0.9901 0.2756 0.9901 0.9950
No log 15.25 366 1.0843 0.2166 1.0843 1.0413
No log 15.3333 368 1.1496 0.2288 1.1496 1.0722
No log 15.4167 370 1.1024 0.2330 1.1024 1.0499
No log 15.5 372 0.9892 0.2781 0.9892 0.9946
No log 15.5833 374 0.8694 0.2012 0.8694 0.9324
No log 15.6667 376 0.8158 0.2409 0.8158 0.9032
No log 15.75 378 0.8676 0.2012 0.8676 0.9315
No log 15.8333 380 0.9813 0.1692 0.9813 0.9906
No log 15.9167 382 1.0579 0.2651 1.0579 1.0285
No log 16.0 384 1.0676 0.2601 1.0676 1.0332
No log 16.0833 386 0.9952 0.2460 0.9952 0.9976
No log 16.1667 388 0.9146 0.2726 0.9146 0.9563
No log 16.25 390 0.8836 0.2518 0.8836 0.9400
No log 16.3333 392 0.8831 0.3606 0.8831 0.9397
No log 16.4167 394 0.9178 0.3433 0.9178 0.9580
No log 16.5 396 1.0362 0.2802 1.0362 1.0179
No log 16.5833 398 1.0937 0.2802 1.0937 1.0458
No log 16.6667 400 0.9775 0.2999 0.9775 0.9887
No log 16.75 402 0.8036 0.2012 0.8036 0.8964
No log 16.8333 404 0.7244 0.2817 0.7244 0.8511
No log 16.9167 406 0.7277 0.2817 0.7277 0.8531
No log 17.0 408 0.7949 0.2817 0.7949 0.8916
No log 17.0833 410 0.7906 0.2817 0.7906 0.8892
No log 17.1667 412 0.7720 0.2817 0.7720 0.8787
No log 17.25 414 0.7766 0.2817 0.7766 0.8813
No log 17.3333 416 0.8542 0.2297 0.8542 0.9242
No log 17.4167 418 0.8541 0.2692 0.8541 0.9242
No log 17.5 420 0.9158 0.3359 0.9158 0.9570
No log 17.5833 422 0.9075 0.2784 0.9075 0.9526
No log 17.6667 424 0.8586 0.2574 0.8586 0.9266
No log 17.75 426 0.8742 0.2193 0.8742 0.9350
No log 17.8333 428 0.9507 0.2046 0.9507 0.9751
No log 17.9167 430 1.0799 0.2756 1.0799 1.0392
No log 18.0 432 1.2145 0.2376 1.2145 1.1020
No log 18.0833 434 1.1990 0.2376 1.1990 1.0950
No log 18.1667 436 1.0312 0.2939 1.0312 1.0155
No log 18.25 438 0.9313 0.2308 0.9313 0.9650
No log 18.3333 440 0.8990 0.2193 0.8990 0.9482
No log 18.4167 442 0.9208 0.2615 0.9208 0.9596
No log 18.5 444 1.0252 0.3516 1.0252 1.0125
No log 18.5833 446 1.1581 0.2247 1.1581 1.0761
No log 18.6667 448 1.1351 0.2545 1.1351 1.0654
No log 18.75 450 1.0173 0.2756 1.0173 1.0086
No log 18.8333 452 0.8774 0.1914 0.8774 0.9367
No log 18.9167 454 0.8164 0.2352 0.8164 0.9036
No log 19.0 456 0.8237 0.2352 0.8237 0.9076
No log 19.0833 458 0.8989 0.1867 0.8989 0.9481
No log 19.1667 460 1.0130 0.2046 1.0130 1.0065
No log 19.25 462 1.0713 0.2297 1.0713 1.0350
No log 19.3333 464 1.0790 0.2948 1.0790 1.0387
No log 19.4167 466 1.0592 0.2881 1.0592 1.0292
No log 19.5 468 0.9846 0.2866 0.9846 0.9922
No log 19.5833 470 0.8960 0.2297 0.8960 0.9465
No log 19.6667 472 0.8743 0.2297 0.8743 0.9351
No log 19.75 474 0.8711 0.2574 0.8711 0.9333
No log 19.8333 476 0.8953 0.3169 0.8953 0.9462
No log 19.9167 478 0.9865 0.3477 0.9865 0.9932
No log 20.0 480 1.0715 0.2926 1.0715 1.0351
No log 20.0833 482 1.0591 0.3359 1.0591 1.0291
No log 20.1667 484 1.0122 0.3417 1.0122 1.0061
No log 20.25 486 0.9730 0.3579 0.9730 0.9864
No log 20.3333 488 0.8519 0.3169 0.8519 0.9230
No log 20.4167 490 0.7732 0.2754 0.7732 0.8793
No log 20.5 492 0.7826 0.2754 0.7826 0.8847
No log 20.5833 494 0.8464 0.2574 0.8464 0.9200
No log 20.6667 496 0.9033 0.2784 0.9033 0.9504
No log 20.75 498 1.0128 0.3477 1.0128 1.0064
0.3113 20.8333 500 1.0712 0.3359 1.0712 1.0350
0.3113 20.9167 502 1.1021 0.3214 1.1021 1.0498
0.3113 21.0 504 1.0396 0.3251 1.0396 1.0196
0.3113 21.0833 506 0.9484 0.2781 0.9484 0.9739
0.3113 21.1667 508 0.9129 0.2784 0.9129 0.9555
0.3113 21.25 510 0.9447 0.2615 0.9447 0.9719

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k10_task7_organization

Finetuned
(4019)
this model