ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k17_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0676
  • Qwk: 0.2683
  • Mse: 1.0676
  • Rmse: 1.0333

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 2.7815 -0.0481 2.7815 1.6678
No log 0.1 4 1.7927 0.0061 1.7927 1.3389
No log 0.15 6 2.0211 -0.1653 2.0211 1.4217
No log 0.2 8 1.3569 -0.1328 1.3569 1.1648
No log 0.25 10 1.0144 0.0054 1.0144 1.0072
No log 0.3 12 0.9010 0.1461 0.9010 0.9492
No log 0.35 14 0.9014 0.1534 0.9014 0.9494
No log 0.4 16 0.8922 0.1636 0.8922 0.9445
No log 0.45 18 0.8487 0.0679 0.8487 0.9212
No log 0.5 20 0.9147 0.1511 0.9147 0.9564
No log 0.55 22 1.0285 0.1259 1.0285 1.0141
No log 0.6 24 1.0813 0.0986 1.0813 1.0398
No log 0.65 26 0.8837 0.2132 0.8837 0.9401
No log 0.7 28 0.7961 0.0937 0.7961 0.8922
No log 0.75 30 0.7689 0.0481 0.7689 0.8769
No log 0.8 32 0.7462 0.0481 0.7462 0.8638
No log 0.85 34 0.7381 0.0884 0.7381 0.8591
No log 0.9 36 0.7520 0.0 0.7520 0.8672
No log 0.95 38 0.7783 0.0481 0.7783 0.8822
No log 1.0 40 0.7709 0.0 0.7709 0.8780
No log 1.05 42 0.7433 0.0 0.7433 0.8622
No log 1.1 44 0.7383 0.0 0.7383 0.8592
No log 1.15 46 0.7397 0.0 0.7397 0.8601
No log 1.2 48 0.7337 0.0884 0.7337 0.8566
No log 1.25 50 0.7301 0.1236 0.7301 0.8544
No log 1.3 52 0.7242 0.1456 0.7242 0.8510
No log 1.35 54 0.7434 0.1807 0.7434 0.8622
No log 1.4 56 0.7364 0.1508 0.7364 0.8581
No log 1.45 58 0.7273 0.1187 0.7273 0.8528
No log 1.5 60 0.7258 0.0840 0.7258 0.8520
No log 1.55 62 0.7335 0.0444 0.7335 0.8564
No log 1.6 64 0.7513 0.0937 0.7513 0.8668
No log 1.65 66 0.7398 0.0481 0.7398 0.8601
No log 1.7 68 0.7442 0.0 0.7442 0.8627
No log 1.75 70 0.7480 0.0 0.7480 0.8649
No log 1.8 72 0.7431 -0.0027 0.7431 0.8620
No log 1.85 74 0.7483 0.0893 0.7483 0.8651
No log 1.9 76 0.7506 0.0026 0.7506 0.8664
No log 1.95 78 0.7455 0.0026 0.7455 0.8634
No log 2.0 80 0.7296 0.0764 0.7296 0.8542
No log 2.05 82 0.7244 0.0410 0.7244 0.8511
No log 2.1 84 0.7185 0.0481 0.7185 0.8476
No log 2.15 86 0.7201 0.0481 0.7201 0.8486
No log 2.2 88 0.7684 0.0688 0.7684 0.8766
No log 2.25 90 0.8471 -0.0047 0.8471 0.9204
No log 2.3 92 0.9219 0.0336 0.9219 0.9602
No log 2.35 94 0.8593 0.0661 0.8593 0.9270
No log 2.4 96 0.7927 0.1448 0.7927 0.8903
No log 2.45 98 0.7396 0.2158 0.7396 0.8600
No log 2.5 100 0.7441 0.2158 0.7441 0.8626
No log 2.55 102 0.7275 0.1867 0.7275 0.8530
No log 2.6 104 0.7325 0.2509 0.7325 0.8559
No log 2.65 106 0.7702 0.2218 0.7702 0.8776
No log 2.7 108 0.7711 0.2158 0.7711 0.8781
No log 2.75 110 0.7585 0.2158 0.7585 0.8709
No log 2.8 112 0.7625 0.2158 0.7625 0.8732
No log 2.85 114 0.7762 0.2413 0.7762 0.8810
No log 2.9 116 0.7775 0.1901 0.7775 0.8818
No log 2.95 118 0.7895 0.2847 0.7895 0.8886
No log 3.0 120 0.7612 0.1624 0.7612 0.8724
No log 3.05 122 0.7445 0.2158 0.7445 0.8629
No log 3.1 124 0.7593 0.1010 0.7593 0.8714
No log 3.15 126 0.8076 0.0971 0.8076 0.8986
No log 3.2 128 0.7975 0.0971 0.7975 0.8930
No log 3.25 130 0.7766 0.0697 0.7766 0.8812
No log 3.3 132 0.7984 0.1051 0.7984 0.8935
No log 3.35 134 0.9101 0.2149 0.9101 0.9540
No log 3.4 136 1.0361 0.2521 1.0361 1.0179
No log 3.45 138 1.0476 0.2364 1.0476 1.0235
No log 3.5 140 0.9632 0.1995 0.9632 0.9814
No log 3.55 142 0.9064 0.0584 0.9064 0.9521
No log 3.6 144 0.8631 0.0697 0.8631 0.9290
No log 3.65 146 0.9331 0.0975 0.9331 0.9660
No log 3.7 148 0.9420 0.0856 0.9420 0.9706
No log 3.75 150 0.9851 0.2193 0.9851 0.9925
No log 3.8 152 0.9770 0.2892 0.9770 0.9885
No log 3.85 154 0.9237 0.2439 0.9237 0.9611
No log 3.9 156 0.8649 0.2943 0.8649 0.9300
No log 3.95 158 0.8627 0.3369 0.8627 0.9288
No log 4.0 160 0.9110 0.2912 0.9110 0.9544
No log 4.05 162 0.8747 0.3115 0.8747 0.9353
No log 4.1 164 0.8485 0.3157 0.8485 0.9211
No log 4.15 166 0.8876 0.2059 0.8876 0.9421
No log 4.2 168 0.8396 0.2662 0.8396 0.9163
No log 4.25 170 0.7145 0.3020 0.7145 0.8453
No log 4.3 172 0.6763 0.1829 0.6763 0.8224
No log 4.35 174 0.6848 0.2181 0.6848 0.8275
No log 4.4 176 0.7393 0.4052 0.7393 0.8598
No log 4.45 178 0.8544 0.4251 0.8544 0.9243
No log 4.5 180 0.8607 0.3754 0.8607 0.9277
No log 4.55 182 0.8286 0.4251 0.8286 0.9103
No log 4.6 184 0.7767 0.3167 0.7767 0.8813
No log 4.65 186 0.7748 0.3167 0.7748 0.8802
No log 4.7 188 0.7632 0.3622 0.7632 0.8736
No log 4.75 190 0.7500 0.3341 0.7500 0.8660
No log 4.8 192 0.7483 0.2950 0.7483 0.8651
No log 4.85 194 0.7769 0.4642 0.7769 0.8814
No log 4.9 196 0.7819 0.5120 0.7819 0.8843
No log 4.95 198 0.8057 0.3789 0.8057 0.8976
No log 5.0 200 0.8061 0.2950 0.8061 0.8978
No log 5.05 202 0.8570 0.2967 0.8570 0.9257
No log 5.1 204 0.8946 0.4462 0.8946 0.9458
No log 5.15 206 0.8111 0.3372 0.8111 0.9006
No log 5.2 208 0.7667 0.2847 0.7667 0.8756
No log 5.25 210 0.8306 0.4247 0.8306 0.9114
No log 5.3 212 0.9020 0.3333 0.9020 0.9498
No log 5.35 214 0.9648 0.3727 0.9648 0.9822
No log 5.4 216 0.9330 0.3012 0.9330 0.9659
No log 5.45 218 0.9431 0.2779 0.9431 0.9712
No log 5.5 220 0.8969 0.1029 0.8969 0.9471
No log 5.55 222 0.8776 0.1918 0.8776 0.9368
No log 5.6 224 0.8238 0.1935 0.8238 0.9076
No log 5.65 226 0.7614 0.3127 0.7614 0.8726
No log 5.7 228 0.7729 0.3399 0.7729 0.8791
No log 5.75 230 0.8301 0.3425 0.8301 0.9111
No log 5.8 232 0.9000 0.3579 0.9000 0.9487
No log 5.85 234 0.9826 0.2886 0.9826 0.9913
No log 5.9 236 1.0048 0.3059 1.0048 1.0024
No log 5.95 238 0.9560 0.2886 0.9560 0.9778
No log 6.0 240 0.9846 0.2886 0.9846 0.9923
No log 6.05 242 1.0053 0.3247 1.0053 1.0026
No log 6.1 244 0.8946 0.2923 0.8946 0.9458
No log 6.15 246 0.8653 0.2518 0.8653 0.9302
No log 6.2 248 0.8387 0.3127 0.8387 0.9158
No log 6.25 250 0.8439 0.3060 0.8439 0.9187
No log 6.3 252 0.8582 0.2632 0.8582 0.9264
No log 6.35 254 0.9131 0.4113 0.9131 0.9556
No log 6.4 256 0.9065 0.4113 0.9065 0.9521
No log 6.45 258 0.8590 0.4462 0.8590 0.9268
No log 6.5 260 0.8227 0.3169 0.8227 0.9070
No log 6.55 262 0.8932 0.3371 0.8932 0.9451
No log 6.6 264 1.0092 0.2802 1.0092 1.0046
No log 6.65 266 1.0151 0.2926 1.0151 1.0075
No log 6.7 268 0.9591 0.3417 0.9591 0.9794
No log 6.75 270 0.9389 0.3579 0.9389 0.9690
No log 6.8 272 0.9699 0.3302 0.9699 0.9848
No log 6.85 274 1.0013 0.2501 1.0013 1.0007
No log 6.9 276 1.0755 0.2264 1.0755 1.0371
No log 6.95 278 1.1285 0.2264 1.1285 1.0623
No log 7.0 280 1.0357 0.2796 1.0357 1.0177
No log 7.05 282 0.9956 0.3608 0.9956 0.9978
No log 7.1 284 0.9900 0.3557 0.9900 0.9950
No log 7.15 286 0.9507 0.3557 0.9507 0.9751
No log 7.2 288 0.8655 0.4862 0.8655 0.9303
No log 7.25 290 0.8401 0.4144 0.8401 0.9166
No log 7.3 292 0.8795 0.4462 0.8795 0.9378
No log 7.35 294 1.0027 0.2732 1.0027 1.0014
No log 7.4 296 1.0642 0.1981 1.0642 1.0316
No log 7.45 298 1.0066 0.3114 1.0066 1.0033
No log 7.5 300 0.8917 0.4541 0.8917 0.9443
No log 7.55 302 0.8794 0.5077 0.8794 0.9378
No log 7.6 304 0.9474 0.4044 0.9474 0.9733
No log 7.65 306 0.9984 0.3031 0.9984 0.9992
No log 7.7 308 0.9133 0.4114 0.9133 0.9557
No log 7.75 310 0.7844 0.2605 0.7844 0.8857
No log 7.8 312 0.7335 0.2809 0.7335 0.8564
No log 7.85 314 0.7312 0.2809 0.7312 0.8551
No log 7.9 316 0.7909 0.3121 0.7909 0.8893
No log 7.95 318 1.0015 0.2389 1.0015 1.0007
No log 8.0 320 1.2296 0.1391 1.2296 1.1089
No log 8.05 322 1.4373 0.1122 1.4373 1.1989
No log 8.1 324 1.4303 0.1122 1.4303 1.1960
No log 8.15 326 1.2038 0.1654 1.2038 1.0972
No log 8.2 328 0.9614 0.2075 0.9614 0.9805
No log 8.25 330 0.8355 0.3564 0.8355 0.9141
No log 8.3 332 0.8083 0.4329 0.8083 0.8991
No log 8.35 334 0.8333 0.4644 0.8333 0.9129
No log 8.4 336 0.8453 0.4627 0.8453 0.9194
No log 8.45 338 0.8036 0.4167 0.8036 0.8965
No log 8.5 340 0.7977 0.3399 0.7977 0.8931
No log 8.55 342 0.8403 0.3544 0.8403 0.9167
No log 8.6 344 0.8785 0.3121 0.8785 0.9373
No log 8.65 346 0.9207 0.3207 0.9207 0.9595
No log 8.7 348 0.9339 0.2669 0.9339 0.9664
No log 8.75 350 0.8966 0.3677 0.8966 0.9469
No log 8.8 352 0.8942 0.3329 0.8942 0.9456
No log 8.85 354 0.9491 0.3560 0.9491 0.9742
No log 8.9 356 1.1092 0.2520 1.1092 1.0532
No log 8.95 358 1.2857 0.2197 1.2857 1.1339
No log 9.0 360 1.3019 0.1793 1.3019 1.1410
No log 9.05 362 1.1466 0.2191 1.1466 1.0708
No log 9.1 364 0.9258 0.4113 0.9258 0.9622
No log 9.15 366 0.8345 0.3746 0.8345 0.9135
No log 9.2 368 0.8240 0.3972 0.8240 0.9077
No log 9.25 370 0.8596 0.4462 0.8596 0.9271
No log 9.3 372 0.9124 0.4008 0.9124 0.9552
No log 9.35 374 0.9869 0.3359 0.9869 0.9934
No log 9.4 376 1.0105 0.2926 1.0105 1.0052
No log 9.45 378 0.9363 0.4328 0.9363 0.9676
No log 9.5 380 0.8247 0.3564 0.8247 0.9081
No log 9.55 382 0.7809 0.2589 0.7809 0.8837
No log 9.6 384 0.7887 0.2589 0.7887 0.8881
No log 9.65 386 0.8558 0.3564 0.8558 0.9251
No log 9.7 388 1.0092 0.2659 1.0092 1.0046
No log 9.75 390 1.1964 0.1805 1.1964 1.0938
No log 9.8 392 1.2204 0.1479 1.2204 1.1047
No log 9.85 394 1.1127 0.2264 1.1127 1.0548
No log 9.9 396 0.9637 0.3516 0.9637 0.9817
No log 9.95 398 0.8570 0.4144 0.8570 0.9257
No log 10.0 400 0.8032 0.3099 0.8032 0.8962
No log 10.05 402 0.8018 0.2527 0.8018 0.8954
No log 10.1 404 0.8452 0.3564 0.8452 0.9193
No log 10.15 406 0.9750 0.2651 0.9750 0.9874
No log 10.2 408 1.1752 0.1508 1.1752 1.0841
No log 10.25 410 1.2968 0.1522 1.2968 1.1388
No log 10.3 412 1.3569 0.1414 1.3569 1.1649
No log 10.35 414 1.2820 0.1549 1.2820 1.1323
No log 10.4 416 1.1277 0.1568 1.1277 1.0619
No log 10.45 418 0.9659 0.3761 0.9659 0.9828
No log 10.5 420 0.8109 0.4247 0.8109 0.9005
No log 10.55 422 0.7443 0.2558 0.7443 0.8628
No log 10.6 424 0.7236 0.2261 0.7236 0.8506
No log 10.65 426 0.7324 0.2261 0.7324 0.8558
No log 10.7 428 0.7808 0.3399 0.7808 0.8836
No log 10.75 430 0.8909 0.4404 0.8909 0.9439
No log 10.8 432 1.0132 0.4044 1.0132 1.0066
No log 10.85 434 1.0463 0.2264 1.0463 1.0229
No log 10.9 436 1.0307 0.1858 1.0307 1.0152
No log 10.95 438 0.9524 0.2510 0.9524 0.9759
No log 11.0 440 0.9381 0.2866 0.9381 0.9686
No log 11.05 442 0.9587 0.2460 0.9587 0.9791
No log 11.1 444 0.9440 0.2510 0.9440 0.9716
No log 11.15 446 0.9761 0.1747 0.9761 0.9880
No log 11.2 448 1.1083 0.1870 1.1083 1.0527
No log 11.25 450 1.1684 0.1203 1.1684 1.0809
No log 11.3 452 1.1929 0.1530 1.1929 1.0922
No log 11.35 454 1.1310 0.1671 1.1310 1.0635
No log 11.4 456 1.0862 0.0922 1.0862 1.0422
No log 11.45 458 1.0900 0.1463 1.0900 1.0440
No log 11.5 460 1.0844 0.1671 1.0844 1.0414
No log 11.55 462 1.0999 0.1909 1.0999 1.0488
No log 11.6 464 1.1240 0.2020 1.1240 1.0602
No log 11.65 466 1.0884 0.2020 1.0884 1.0433
No log 11.7 468 1.0469 0.2732 1.0469 1.0232
No log 11.75 470 1.0083 0.2271 1.0083 1.0041
No log 11.8 472 0.9795 0.2119 0.9795 0.9897
No log 11.85 474 0.9908 0.2703 0.9908 0.9954
No log 11.9 476 1.0087 0.2886 1.0087 1.0044
No log 11.95 478 1.0151 0.2552 1.0151 1.0075
No log 12.0 480 1.0537 0.2045 1.0537 1.0265
No log 12.05 482 1.0959 0.2006 1.0959 1.0468
No log 12.1 484 1.0750 0.2006 1.0750 1.0368
No log 12.15 486 1.0557 0.2833 1.0557 1.0275
No log 12.2 488 1.0304 0.2886 1.0304 1.0151
No log 12.25 490 1.0110 0.2995 1.0110 1.0055
No log 12.3 492 0.9812 0.3110 0.9812 0.9906
No log 12.35 494 0.9318 0.3606 0.9318 0.9653
No log 12.4 496 0.9165 0.3359 0.9165 0.9573
No log 12.45 498 0.9682 0.2487 0.9682 0.9840
0.3594 12.5 500 1.0243 0.3193 1.0243 1.0121
0.3594 12.55 502 0.9971 0.3193 0.9971 0.9985
0.3594 12.6 504 0.9262 0.3846 0.9262 0.9624
0.3594 12.65 506 0.9074 0.3918 0.9074 0.9526
0.3594 12.7 508 0.9062 0.3991 0.9062 0.9520
0.3594 12.75 510 0.9505 0.2703 0.9505 0.9750
0.3594 12.8 512 1.0398 0.2683 1.0398 1.0197
0.3594 12.85 514 1.1322 0.2059 1.1322 1.0640
0.3594 12.9 516 1.1282 0.2059 1.1282 1.0622
0.3594 12.95 518 1.0676 0.2683 1.0676 1.0333

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k17_task7_organization

Finetuned
(4019)
this model