ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k9_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8619
  • Qwk: -0.0766
  • Mse: 0.8619
  • Rmse: 0.9284

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0870 2 3.6497 0.0035 3.6497 1.9104
No log 0.1739 4 2.0566 0.0454 2.0566 1.4341
No log 0.2609 6 1.5936 -0.0041 1.5936 1.2624
No log 0.3478 8 0.9922 0.0157 0.9922 0.9961
No log 0.4348 10 0.9854 -0.0923 0.9854 0.9927
No log 0.5217 12 1.4309 -0.1001 1.4309 1.1962
No log 0.6087 14 1.2876 -0.0751 1.2876 1.1347
No log 0.6957 16 0.8814 0.1007 0.8814 0.9388
No log 0.7826 18 0.9140 0.0819 0.9140 0.9560
No log 0.8696 20 1.0444 0.0016 1.0444 1.0220
No log 0.9565 22 1.0496 0.0 1.0496 1.0245
No log 1.0435 24 1.0389 0.0 1.0389 1.0193
No log 1.1304 26 1.3096 0.0 1.3096 1.1444
No log 1.2174 28 1.5050 0.0 1.5050 1.2268
No log 1.3043 30 1.3023 0.0 1.3023 1.1412
No log 1.3913 32 1.0549 0.0048 1.0549 1.0271
No log 1.4783 34 0.8689 -0.1676 0.8689 0.9321
No log 1.5652 36 0.8100 -0.1730 0.8100 0.9000
No log 1.6522 38 0.8260 -0.1722 0.8260 0.9089
No log 1.7391 40 1.0154 0.0518 1.0154 1.0077
No log 1.8261 42 1.2994 0.0 1.2994 1.1399
No log 1.9130 44 1.2533 -0.0234 1.2533 1.1195
No log 2.0 46 1.1819 -0.0479 1.1819 1.0871
No log 2.0870 48 1.1236 -0.0457 1.1236 1.0600
No log 2.1739 50 1.0592 -0.0178 1.0592 1.0292
No log 2.2609 52 1.1286 -0.0207 1.1286 1.0623
No log 2.3478 54 1.1331 -0.0207 1.1331 1.0645
No log 2.4348 56 1.0483 -0.0164 1.0483 1.0239
No log 2.5217 58 0.9950 -0.0149 0.9950 0.9975
No log 2.6087 60 1.0374 0.0100 1.0374 1.0185
No log 2.6957 62 0.9737 0.0543 0.9737 0.9868
No log 2.7826 64 0.9036 -0.0504 0.9036 0.9506
No log 2.8696 66 1.0664 0.0176 1.0664 1.0327
No log 2.9565 68 1.4343 -0.0479 1.4343 1.1976
No log 3.0435 70 1.3352 -0.0468 1.3352 1.1555
No log 3.1304 72 1.0813 -0.0972 1.0813 1.0399
No log 3.2174 74 0.9358 0.0233 0.9358 0.9674
No log 3.3043 76 0.9447 0.0805 0.9447 0.9719
No log 3.3913 78 0.9375 0.0515 0.9375 0.9682
No log 3.4783 80 1.0077 -0.0912 1.0077 1.0039
No log 3.5652 82 0.9976 0.0129 0.9976 0.9988
No log 3.6522 84 0.8169 -0.0739 0.8169 0.9038
No log 3.7391 86 0.7513 -0.0035 0.7513 0.8668
No log 3.8261 88 0.7713 0.0374 0.7713 0.8782
No log 3.9130 90 0.9667 -0.0459 0.9667 0.9832
No log 4.0 92 1.3463 -0.0123 1.3463 1.1603
No log 4.0870 94 1.0481 0.0526 1.0481 1.0237
No log 4.1739 96 0.7866 -0.0125 0.7866 0.8869
No log 4.2609 98 0.8419 -0.0345 0.8419 0.9175
No log 4.3478 100 0.7959 -0.0033 0.7959 0.8921
No log 4.4348 102 0.9685 -0.0373 0.9685 0.9841
No log 4.5217 104 1.4208 -0.1243 1.4208 1.1920
No log 4.6087 106 1.6191 -0.1729 1.6191 1.2724
No log 4.6957 108 1.2961 -0.1905 1.2961 1.1385
No log 4.7826 110 0.9342 -0.0718 0.9342 0.9666
No log 4.8696 112 0.8826 -0.1094 0.8826 0.9395
No log 4.9565 114 0.8711 -0.1158 0.8711 0.9333
No log 5.0435 116 0.8878 -0.0228 0.8878 0.9423
No log 5.1304 118 0.9042 0.0438 0.9042 0.9509
No log 5.2174 120 0.8299 0.0574 0.8299 0.9110
No log 5.3043 122 0.7751 -0.0550 0.7751 0.8804
No log 5.3913 124 0.7915 -0.1468 0.7915 0.8897
No log 5.4783 126 0.7954 0.0289 0.7954 0.8919
No log 5.5652 128 1.0492 0.0912 1.0492 1.0243
No log 5.6522 130 1.0253 0.0949 1.0253 1.0126
No log 5.7391 132 0.8117 -0.0643 0.8117 0.9010
No log 5.8261 134 0.8318 -0.1268 0.8318 0.9120
No log 5.9130 136 0.8243 -0.1978 0.8243 0.9079
No log 6.0 138 0.8298 -0.0675 0.8298 0.9109
No log 6.0870 140 1.0118 0.0333 1.0118 1.0059
No log 6.1739 142 0.9890 -0.0008 0.9890 0.9945
No log 6.2609 144 0.8455 -0.1131 0.8455 0.9195
No log 6.3478 146 0.7751 -0.0541 0.7751 0.8804
No log 6.4348 148 0.7887 0.0602 0.7887 0.8881
No log 6.5217 150 0.7604 -0.0560 0.7604 0.8720
No log 6.6087 152 0.9895 -0.0008 0.9895 0.9947
No log 6.6957 154 1.1260 -0.1181 1.1260 1.0611
No log 6.7826 156 0.9206 0.0091 0.9206 0.9595
No log 6.8696 158 0.7933 -0.1100 0.7933 0.8907
No log 6.9565 160 0.8030 -0.1538 0.8030 0.8961
No log 7.0435 162 0.8281 -0.0690 0.8281 0.9100
No log 7.1304 164 0.9789 -0.0351 0.9789 0.9894
No log 7.2174 166 1.0475 -0.0797 1.0475 1.0235
No log 7.3043 168 0.9485 0.0504 0.9485 0.9739
No log 7.3913 170 0.8061 0.0432 0.8061 0.8978
No log 7.4783 172 0.8110 -0.0385 0.8110 0.9005
No log 7.5652 174 0.7632 -0.0428 0.7632 0.8736
No log 7.6522 176 0.7445 -0.0091 0.7445 0.8628
No log 7.7391 178 0.7937 0.0183 0.7937 0.8909
No log 7.8261 180 0.7374 0.0814 0.7374 0.8587
No log 7.9130 182 0.7034 -0.0033 0.7034 0.8387
No log 8.0 184 0.7136 0.0964 0.7136 0.8447
No log 8.0870 186 0.7487 0.1379 0.7487 0.8653
No log 8.1739 188 0.7340 0.0964 0.7340 0.8568
No log 8.2609 190 0.7558 0.1254 0.7558 0.8694
No log 8.3478 192 0.8471 0.0041 0.8471 0.9204
No log 8.4348 194 0.9546 0.0348 0.9546 0.9770
No log 8.5217 196 0.8747 0.0043 0.8747 0.9353
No log 8.6087 198 0.8720 -0.0351 0.8720 0.9338
No log 8.6957 200 0.9089 -0.0371 0.9089 0.9533
No log 8.7826 202 0.8373 0.0159 0.8373 0.9151
No log 8.8696 204 0.7852 -0.0215 0.7852 0.8861
No log 8.9565 206 0.7877 -0.0560 0.7877 0.8875
No log 9.0435 208 0.8023 -0.0560 0.8023 0.8957
No log 9.1304 210 0.8168 -0.1106 0.8168 0.9038
No log 9.2174 212 0.8539 -0.1083 0.8539 0.9241
No log 9.3043 214 0.9933 0.0392 0.9933 0.9967
No log 9.3913 216 0.9901 -0.0050 0.9901 0.9950
No log 9.4783 218 0.8379 -0.0686 0.8379 0.9154
No log 9.5652 220 0.7845 -0.0627 0.7845 0.8857
No log 9.6522 222 0.7677 -0.0125 0.7677 0.8762
No log 9.7391 224 0.7955 -0.0264 0.7955 0.8919
No log 9.8261 226 0.8608 0.0043 0.8608 0.9278
No log 9.9130 228 0.9484 0.0346 0.9484 0.9739
No log 10.0 230 0.8713 -0.0743 0.8713 0.9334
No log 10.0870 232 0.8146 -0.2077 0.8146 0.9026
No log 10.1739 234 0.8216 -0.1026 0.8216 0.9064
No log 10.2609 236 0.8264 -0.2109 0.8264 0.9091
No log 10.3478 238 0.8616 -0.0731 0.8616 0.9282
No log 10.4348 240 0.9414 -0.0355 0.9414 0.9703
No log 10.5217 242 0.8998 -0.0331 0.8998 0.9486
No log 10.6087 244 0.8216 -0.0287 0.8216 0.9064
No log 10.6957 246 0.7843 -0.2183 0.7843 0.8856
No log 10.7826 248 0.7790 -0.0660 0.7790 0.8826
No log 10.8696 250 0.7794 -0.1668 0.7794 0.8828
No log 10.9565 252 0.8316 -0.0731 0.8316 0.9119
No log 11.0435 254 0.9614 0.0392 0.9614 0.9805
No log 11.1304 256 0.9304 0.0016 0.9304 0.9646
No log 11.2174 258 0.8093 -0.0170 0.8093 0.8996
No log 11.3043 260 0.8113 -0.1026 0.8113 0.9007
No log 11.3913 262 0.8426 0.0196 0.8426 0.9179
No log 11.4783 264 0.8199 -0.0428 0.8199 0.9055
No log 11.5652 266 0.7910 -0.0091 0.7910 0.8894
No log 11.6522 268 0.8258 -0.0704 0.8258 0.9087
No log 11.7391 270 0.7937 -0.0215 0.7937 0.8909
No log 11.8261 272 0.7623 -0.0125 0.7623 0.8731
No log 11.9130 274 0.7577 -0.0065 0.7577 0.8705
No log 12.0 276 0.7599 -0.0591 0.7599 0.8717
No log 12.0870 278 0.7770 -0.0679 0.7770 0.8815
No log 12.1739 280 0.8348 -0.0215 0.8348 0.9137
No log 12.2609 282 0.8826 -0.0753 0.8826 0.9395
No log 12.3478 284 0.9215 -0.0699 0.9215 0.9599
No log 12.4348 286 0.8545 -0.1168 0.8545 0.9244
No log 12.5217 288 0.8436 0.0031 0.8436 0.9185
No log 12.6087 290 0.8370 -0.1168 0.8370 0.9149
No log 12.6957 292 0.8268 -0.0718 0.8268 0.9093
No log 12.7826 294 0.7802 -0.0739 0.7802 0.8833
No log 12.8696 296 0.7556 -0.0188 0.7556 0.8693
No log 12.9565 298 0.7632 -0.0240 0.7632 0.8736
No log 13.0435 300 0.8258 -0.0309 0.8258 0.9087
No log 13.1304 302 0.8489 -0.0309 0.8489 0.9214
No log 13.2174 304 0.8039 -0.0309 0.8039 0.8966
No log 13.3043 306 0.7960 -0.0240 0.7960 0.8922
No log 13.3913 308 0.7891 -0.0711 0.7891 0.8883
No log 13.4783 310 0.8520 -0.0240 0.8520 0.9230
No log 13.5652 312 0.9489 -0.0331 0.9489 0.9741
No log 13.6522 314 1.0388 -0.0456 1.0388 1.0192
No log 13.7391 316 0.9796 -0.0801 0.9796 0.9897
No log 13.8261 318 0.8848 -0.0264 0.8848 0.9406
No log 13.9130 320 0.8072 -0.0675 0.8072 0.8984
No log 14.0 322 0.8037 -0.0690 0.8037 0.8965
No log 14.0870 324 0.8251 -0.0264 0.8251 0.9084
No log 14.1739 326 0.8962 0.0071 0.8962 0.9467
No log 14.2609 328 0.9566 -0.0828 0.9566 0.9781
No log 14.3478 330 0.8937 0.0424 0.8937 0.9454
No log 14.4348 332 0.7787 -0.0240 0.7787 0.8824
No log 14.5217 334 0.7326 -0.0033 0.7326 0.8559
No log 14.6087 336 0.7782 -0.0033 0.7782 0.8822
No log 14.6957 338 0.7741 -0.0033 0.7741 0.8798
No log 14.7826 340 0.7693 -0.0033 0.7693 0.8771
No log 14.8696 342 0.8041 0.0303 0.8041 0.8967
No log 14.9565 344 0.8380 -0.0240 0.8380 0.9154
No log 15.0435 346 0.9189 -0.0723 0.9189 0.9586
No log 15.1304 348 1.0009 -0.1148 1.0009 1.0005
No log 15.2174 350 0.9597 -0.0743 0.9597 0.9796
No log 15.3043 352 0.8619 -0.0588 0.8619 0.9284
No log 15.3913 354 0.8514 0.0376 0.8514 0.9227
No log 15.4783 356 0.8589 0.0289 0.8589 0.9268
No log 15.5652 358 0.8169 0.1249 0.8169 0.9038
No log 15.6522 360 0.7933 0.0821 0.7933 0.8907
No log 15.7391 362 0.7789 0.0414 0.7789 0.8825
No log 15.8261 364 0.7912 0.0282 0.7912 0.8895
No log 15.9130 366 0.7952 -0.0228 0.7952 0.8917
No log 16.0 368 0.7701 0.1371 0.7701 0.8775
No log 16.0870 370 0.7888 0.1371 0.7888 0.8881
No log 16.1739 372 0.7921 0.1371 0.7921 0.8900
No log 16.2609 374 0.7724 0.0857 0.7724 0.8789
No log 16.3478 376 0.7655 -0.0240 0.7655 0.8749
No log 16.4348 378 0.7746 -0.0240 0.7746 0.8801
No log 16.5217 380 0.9044 0.0476 0.9044 0.9510
No log 16.6087 382 1.0211 -0.0886 1.0211 1.0105
No log 16.6957 384 0.9581 -0.0056 0.9581 0.9788
No log 16.7826 386 0.8020 0.0225 0.8020 0.8955
No log 16.8696 388 0.7370 0.1567 0.7370 0.8585
No log 16.9565 390 0.7423 0.0541 0.7423 0.8616
No log 17.0435 392 0.7383 -0.0033 0.7383 0.8592
No log 17.1304 394 0.7230 -0.0033 0.7230 0.8503
No log 17.2174 396 0.7347 0.0909 0.7347 0.8571
No log 17.3043 398 0.7546 0.0395 0.7546 0.8687
No log 17.3913 400 0.7592 0.0395 0.7592 0.8713
No log 17.4783 402 0.7728 0.0918 0.7728 0.8791
No log 17.5652 404 0.7999 0.0085 0.7999 0.8944
No log 17.6522 406 0.8195 0.0583 0.8195 0.9053
No log 17.7391 408 0.7986 0.0488 0.7986 0.8937
No log 17.8261 410 0.7783 0.0355 0.7783 0.8822
No log 17.9130 412 0.8303 0.0159 0.8303 0.9112
No log 18.0 414 0.9833 0.0409 0.9833 0.9916
No log 18.0870 416 1.1352 0.0134 1.1352 1.0655
No log 18.1739 418 1.0528 -0.0878 1.0528 1.0261
No log 18.2609 420 0.8803 -0.0351 0.8803 0.9382
No log 18.3478 422 0.8176 -0.0287 0.8176 0.9042
No log 18.4348 424 0.7689 -0.0125 0.7689 0.8769
No log 18.5217 426 0.7674 -0.0125 0.7674 0.8760
No log 18.6087 428 0.7722 -0.0125 0.7722 0.8788
No log 18.6957 430 0.7532 -0.0096 0.7532 0.8679
No log 18.7826 432 0.7510 0.0374 0.7510 0.8666
No log 18.8696 434 0.7609 0.0714 0.7609 0.8723
No log 18.9565 436 0.7387 0.0807 0.7387 0.8595
No log 19.0435 438 0.7483 0.0296 0.7483 0.8650
No log 19.1304 440 0.7711 0.0260 0.7711 0.8781
No log 19.2174 442 0.7856 0.0260 0.7856 0.8863
No log 19.3043 444 0.7763 0.0260 0.7763 0.8811
No log 19.3913 446 0.7677 -0.0096 0.7677 0.8762
No log 19.4783 448 0.7772 -0.0609 0.7772 0.8816
No log 19.5652 450 0.7977 -0.0560 0.7977 0.8931
No log 19.6522 452 0.8193 0.0031 0.8193 0.9051
No log 19.7391 454 0.8319 0.0061 0.8319 0.9121
No log 19.8261 456 0.8276 -0.0096 0.8276 0.9097
No log 19.9130 458 0.8247 0.0395 0.8247 0.9081
No log 20.0 460 0.8318 -0.0179 0.8318 0.9120
No log 20.0870 462 0.8753 -0.0766 0.8753 0.9356
No log 20.1739 464 0.8578 -0.0766 0.8578 0.9262
No log 20.2609 466 0.8217 0.0260 0.8217 0.9065
No log 20.3478 468 0.8073 0.0296 0.8073 0.8985
No log 20.4348 470 0.7858 -0.0101 0.7858 0.8865
No log 20.5217 472 0.7741 -0.0101 0.7741 0.8798
No log 20.6087 474 0.7640 -0.0101 0.7640 0.8741
No log 20.6957 476 0.7643 0.0416 0.7643 0.8742
No log 20.7826 478 0.7964 0.0260 0.7964 0.8924
No log 20.8696 480 0.8511 -0.0331 0.8511 0.9225
No log 20.9565 482 0.7968 -0.0766 0.7968 0.8926
No log 21.0435 484 0.7347 0.0416 0.7347 0.8571
No log 21.1304 486 0.7201 -0.0101 0.7201 0.8486
No log 21.2174 488 0.7379 0.0416 0.7379 0.8590
No log 21.3043 490 0.7627 0.0334 0.7627 0.8733
No log 21.3913 492 0.8085 0.0260 0.8085 0.8991
No log 21.4783 494 0.8363 -0.0766 0.8363 0.9145
No log 21.5652 496 0.8348 -0.0753 0.8348 0.9137
No log 21.6522 498 0.8119 0.0260 0.8119 0.9011
0.2784 21.7391 500 0.7853 0.0296 0.7853 0.8862
0.2784 21.8261 502 0.7677 -0.0188 0.7677 0.8762
0.2784 21.9130 504 0.7618 -0.0188 0.7618 0.8728
0.2784 22.0 506 0.7464 0.0374 0.7464 0.8639
0.2784 22.0870 508 0.7395 0.0374 0.7395 0.8599
0.2784 22.1739 510 0.7395 0.0374 0.7395 0.8600
0.2784 22.2609 512 0.7451 0.0374 0.7451 0.8632
0.2784 22.3478 514 0.7635 -0.0188 0.7635 0.8738
0.2784 22.4348 516 0.8118 0.0225 0.8118 0.9010
0.2784 22.5217 518 0.8530 -0.0778 0.8530 0.9236
0.2784 22.6087 520 0.8713 -0.0766 0.8713 0.9334
0.2784 22.6957 522 0.8619 -0.0766 0.8619 0.9284

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k9_task3_organization

Finetuned
(4019)
this model