ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k13_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1205
  • Qwk: -0.0961
  • Mse: 1.1205
  • Rmse: 1.0586

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0606 2 3.8797 0.0017 3.8797 1.9697
No log 0.1212 4 2.2173 0.0672 2.2173 1.4891
No log 0.1818 6 1.3323 0.0 1.3323 1.1543
No log 0.2424 8 1.8831 0.0425 1.8831 1.3723
No log 0.3030 10 1.7085 0.0 1.7085 1.3071
No log 0.3636 12 1.0686 -0.0457 1.0686 1.0337
No log 0.4242 14 0.7419 -0.1227 0.7419 0.8613
No log 0.4848 16 0.7259 -0.0069 0.7259 0.8520
No log 0.5455 18 0.7743 -0.1230 0.7743 0.8800
No log 0.6061 20 0.9668 0.0423 0.9668 0.9833
No log 0.6667 22 1.1082 -0.0247 1.1082 1.0527
No log 0.7273 24 1.2468 0.0 1.2468 1.1166
No log 0.7879 26 1.4488 0.0 1.4488 1.2036
No log 0.8485 28 1.4140 0.0 1.4140 1.1891
No log 0.9091 30 1.2783 0.0 1.2783 1.1306
No log 0.9697 32 1.0355 -0.0490 1.0355 1.0176
No log 1.0303 34 1.1671 -0.0490 1.1671 1.0803
No log 1.0909 36 1.2888 -0.0247 1.2888 1.1352
No log 1.1515 38 1.0918 -0.0728 1.0918 1.0449
No log 1.2121 40 1.0713 -0.0728 1.0713 1.0350
No log 1.2727 42 1.2809 -0.0751 1.2809 1.1318
No log 1.3333 44 1.4192 -0.0234 1.4192 1.1913
No log 1.3939 46 1.2224 -0.0728 1.2224 1.1056
No log 1.4545 48 0.9249 -0.0425 0.9249 0.9617
No log 1.5152 50 0.9596 -0.1579 0.9596 0.9796
No log 1.5758 52 0.9341 -0.1586 0.9341 0.9665
No log 1.6364 54 1.0390 -0.0948 1.0390 1.0193
No log 1.6970 56 1.3858 -0.0234 1.3858 1.1772
No log 1.7576 58 1.6152 0.0 1.6152 1.2709
No log 1.8182 60 1.3975 -0.0247 1.3975 1.1822
No log 1.8788 62 0.9126 -0.0456 0.9126 0.9553
No log 1.9394 64 0.7597 -0.0541 0.7597 0.8716
No log 2.0 66 0.7675 -0.0560 0.7675 0.8761
No log 2.0606 68 1.0588 0.0006 1.0588 1.0290
No log 2.1212 70 1.2690 -0.0692 1.2690 1.1265
No log 2.1818 72 1.1748 -0.0677 1.1748 1.0839
No log 2.2424 74 0.9064 -0.0054 0.9064 0.9520
No log 2.3030 76 0.8812 0.0287 0.8812 0.9387
No log 2.3636 78 0.9853 -0.0218 0.9853 0.9926
No log 2.4242 80 1.0275 -0.0236 1.0275 1.0136
No log 2.4848 82 0.8731 -0.0731 0.8731 0.9344
No log 2.5455 84 0.8264 -0.1463 0.8264 0.9091
No log 2.6061 86 0.8607 -0.0984 0.8607 0.9278
No log 2.6667 88 1.0072 -0.1263 1.0072 1.0036
No log 2.7273 90 0.9248 -0.1795 0.9248 0.9617
No log 2.7879 92 0.9311 -0.1466 0.9311 0.9649
No log 2.8485 94 1.0183 -0.1509 1.0183 1.0091
No log 2.9091 96 0.9124 -0.2036 0.9124 0.9552
No log 2.9697 98 0.8933 -0.0939 0.8933 0.9451
No log 3.0303 100 0.8449 -0.2051 0.8449 0.9192
No log 3.0909 102 1.1001 -0.0902 1.1001 1.0489
No log 3.1515 104 0.9967 -0.1253 0.9967 0.9984
No log 3.2121 106 0.8358 -0.1001 0.8358 0.9142
No log 3.2727 108 0.8828 -0.1399 0.8828 0.9396
No log 3.3333 110 0.8588 -0.1001 0.8588 0.9267
No log 3.3939 112 0.8501 -0.1106 0.8501 0.9220
No log 3.4545 114 1.4038 -0.0114 1.4038 1.1848
No log 3.5152 116 1.5872 -0.0445 1.5872 1.2598
No log 3.5758 118 1.0194 -0.1166 1.0194 1.0097
No log 3.6364 120 0.8649 -0.1033 0.8649 0.9300
No log 3.6970 122 0.8876 -0.2008 0.8876 0.9421
No log 3.7576 124 1.1327 -0.0500 1.1327 1.0643
No log 3.8182 126 1.2480 -0.0961 1.2480 1.1171
No log 3.8788 128 0.9311 -0.0336 0.9311 0.9649
No log 3.9394 130 0.9125 -0.2274 0.9125 0.9553
No log 4.0 132 0.9565 -0.0685 0.9565 0.9780
No log 4.0606 134 0.8051 -0.0902 0.8051 0.8973
No log 4.1212 136 1.0921 0.0353 1.0921 1.0450
No log 4.1818 138 1.4618 -0.0174 1.4618 1.2090
No log 4.2424 140 0.9575 0.0576 0.9575 0.9785
No log 4.3030 142 0.8271 -0.1331 0.8271 0.9094
No log 4.3636 144 0.9153 -0.0121 0.9153 0.9567
No log 4.4242 146 0.8253 -0.0449 0.8253 0.9084
No log 4.4848 148 1.0117 -0.1166 1.0117 1.0058
No log 4.5455 150 1.1746 0.0391 1.1746 1.0838
No log 4.6061 152 0.9457 -0.1162 0.9457 0.9725
No log 4.6667 154 0.7764 -0.1033 0.7764 0.8811
No log 4.7273 156 0.7782 -0.0449 0.7782 0.8821
No log 4.7879 158 0.7544 -0.0541 0.7544 0.8686
No log 4.8485 160 0.7402 0.0807 0.7402 0.8603
No log 4.9091 162 0.7756 0.0670 0.7756 0.8807
No log 4.9697 164 0.8344 -0.0833 0.8344 0.9135
No log 5.0303 166 0.8136 0.0449 0.8136 0.9020
No log 5.0909 168 0.8440 -0.0831 0.8440 0.9187
No log 5.1515 170 0.8211 -0.0138 0.8211 0.9062
No log 5.2121 172 0.7743 -0.1470 0.7743 0.8799
No log 5.2727 174 0.7393 -0.0131 0.7393 0.8598
No log 5.3333 176 0.7175 0.0460 0.7175 0.8471
No log 5.3939 178 0.7171 -0.0591 0.7171 0.8468
No log 5.4545 180 0.7598 -0.1018 0.7598 0.8717
No log 5.5152 182 0.8571 -0.0173 0.8571 0.9258
No log 5.5758 184 1.0699 0.0541 1.0699 1.0344
No log 5.6364 186 1.0653 0.0196 1.0653 1.0321
No log 5.6970 188 1.0510 -0.0099 1.0510 1.0252
No log 5.7576 190 0.9058 0.1163 0.9058 0.9517
No log 5.8182 192 0.7647 -0.1018 0.7647 0.8745
No log 5.8788 194 0.7565 -0.0499 0.7565 0.8698
No log 5.9394 196 0.7456 -0.0499 0.7456 0.8635
No log 6.0 198 0.7847 -0.1001 0.7847 0.8859
No log 6.0606 200 0.9803 -0.0892 0.9803 0.9901
No log 6.1212 202 1.0984 0.0859 1.0984 1.0480
No log 6.1818 204 0.8555 0.1458 0.8555 0.9249
No log 6.2424 206 0.8533 0.0456 0.8533 0.9238
No log 6.3030 208 0.9036 -0.0828 0.9036 0.9506
No log 6.3636 210 0.9742 -0.0490 0.9742 0.9870
No log 6.4242 212 0.7867 0.0863 0.7867 0.8870
No log 6.4848 214 0.8560 -0.1833 0.8560 0.9252
No log 6.5455 216 0.8612 -0.0469 0.8612 0.9280
No log 6.6061 218 0.8980 -0.1905 0.8980 0.9476
No log 6.6667 220 1.0304 -0.0861 1.0304 1.0151
No log 6.7273 222 0.9000 -0.1191 0.9000 0.9487
No log 6.7879 224 0.8055 -0.0499 0.8055 0.8975
No log 6.8485 226 0.8017 -0.0499 0.8017 0.8954
No log 6.9091 228 0.8555 -0.0406 0.8555 0.9249
No log 6.9697 230 0.8026 -0.0551 0.8026 0.8959
No log 7.0303 232 0.7640 -0.0551 0.7640 0.8741
No log 7.0909 234 0.7813 -0.0695 0.7813 0.8839
No log 7.1515 236 0.9126 -0.0833 0.9126 0.9553
No log 7.2121 238 0.7967 -0.0695 0.7967 0.8926
No log 7.2727 240 0.8069 -0.1001 0.8069 0.8983
No log 7.3333 242 0.8503 -0.0406 0.8503 0.9221
No log 7.3939 244 0.7919 -0.1001 0.7919 0.8899
No log 7.4545 246 0.7783 -0.0609 0.7783 0.8822
No log 7.5152 248 0.9838 -0.0861 0.9838 0.9918
No log 7.5758 250 0.9866 -0.0474 0.9866 0.9933
No log 7.6364 252 0.7813 -0.0711 0.7813 0.8839
No log 7.6970 254 0.8102 -0.0499 0.8102 0.9001
No log 7.7576 256 0.9412 -0.0377 0.9412 0.9702
No log 7.8182 258 0.9315 -0.0801 0.9315 0.9652
No log 7.8788 260 0.8359 -0.1535 0.8359 0.9143
No log 7.9394 262 0.9858 -0.0809 0.9858 0.9929
No log 8.0 264 0.9722 -0.0442 0.9722 0.9860
No log 8.0606 266 0.8134 -0.0675 0.8134 0.9019
No log 8.1212 268 0.7721 -0.1001 0.7721 0.8787
No log 8.1818 270 0.7534 -0.0551 0.7534 0.8680
No log 8.2424 272 0.7765 -0.1163 0.7765 0.8812
No log 8.3030 274 0.8458 -0.0755 0.8458 0.9197
No log 8.3636 276 0.8814 -0.1614 0.8814 0.9389
No log 8.4242 278 0.7991 -0.0032 0.7991 0.8939
No log 8.4848 280 0.9429 0.0030 0.9429 0.9710
No log 8.5455 282 1.0822 -0.0337 1.0822 1.0403
No log 8.6061 284 0.9868 -0.0187 0.9868 0.9934
No log 8.6667 286 0.8365 0.0030 0.8365 0.9146
No log 8.7273 288 0.8284 0.0776 0.8284 0.9101
No log 8.7879 290 0.7898 0.0030 0.7898 0.8887
No log 8.8485 292 0.7814 -0.0949 0.7814 0.8839
No log 8.9091 294 0.8078 -0.0406 0.8078 0.8988
No log 8.9697 296 0.7851 -0.0473 0.7851 0.8861
No log 9.0303 298 0.7716 0.0375 0.7716 0.8784
No log 9.0909 300 0.7952 -0.0743 0.7952 0.8917
No log 9.1515 302 0.7944 -0.0252 0.7944 0.8913
No log 9.2121 304 0.8055 -0.1074 0.8055 0.8975
No log 9.2727 306 0.8837 -0.0934 0.8837 0.9400
No log 9.3333 308 0.8552 -0.0214 0.8552 0.9248
No log 9.3939 310 0.8180 0.0930 0.8180 0.9044
No log 9.4545 312 0.8489 0.0165 0.8489 0.9213
No log 9.5152 314 0.8834 -0.0513 0.8834 0.9399
No log 9.5758 316 0.8405 -0.0029 0.8405 0.9168
No log 9.6364 318 0.8414 -0.0274 0.8414 0.9173
No log 9.6970 320 0.8706 -0.0316 0.8706 0.9331
No log 9.7576 322 0.8597 -0.1066 0.8597 0.9272
No log 9.8182 324 0.8640 -0.0316 0.8640 0.9295
No log 9.8788 326 0.8177 -0.1168 0.8177 0.9043
No log 9.9394 328 0.8018 -0.1158 0.8018 0.8954
No log 10.0 330 0.8079 -0.1158 0.8079 0.8988
No log 10.0606 332 0.8285 -0.0499 0.8285 0.9102
No log 10.1212 334 0.9354 -0.1475 0.9354 0.9672
No log 10.1818 336 0.9464 -0.1011 0.9464 0.9728
No log 10.2424 338 1.0141 -0.0551 1.0141 1.0070
No log 10.3030 340 0.9885 -0.0616 0.9885 0.9943
No log 10.3636 342 0.8971 -0.1106 0.8971 0.9472
No log 10.4242 344 0.9418 -0.0778 0.9418 0.9704
No log 10.4848 346 0.8766 -0.1168 0.8766 0.9363
No log 10.5455 348 0.8518 -0.2036 0.8518 0.9229
No log 10.6061 350 0.8703 -0.1468 0.8703 0.9329
No log 10.6667 352 0.8924 -0.1527 0.8924 0.9447
No log 10.7273 354 0.9948 -0.0818 0.9948 0.9974
No log 10.7879 356 1.1457 -0.0955 1.1457 1.0704
No log 10.8485 358 1.0569 -0.0253 1.0569 1.0281
No log 10.9091 360 0.9004 -0.0228 0.9004 0.9489
No log 10.9697 362 1.0126 -0.2131 1.0126 1.0063
No log 11.0303 364 1.0450 -0.1342 1.0450 1.0223
No log 11.0909 366 0.9786 -0.1422 0.9786 0.9893
No log 11.1515 368 0.9413 -0.0970 0.9413 0.9702
No log 11.2121 370 0.9375 -0.0316 0.9375 0.9682
No log 11.2727 372 0.9279 -0.1951 0.9279 0.9633
No log 11.3333 374 0.9142 -0.1086 0.9142 0.9561
No log 11.3939 376 0.8692 -0.0939 0.8692 0.9323
No log 11.4545 378 0.8874 -0.1172 0.8874 0.9420
No log 11.5152 380 0.9203 -0.1180 0.9203 0.9593
No log 11.5758 382 0.9359 -0.2022 0.9359 0.9674
No log 11.6364 384 0.9597 -0.1939 0.9597 0.9796
No log 11.6970 386 0.9824 -0.2720 0.9824 0.9911
No log 11.7576 388 0.9770 -0.0860 0.9770 0.9884
No log 11.8182 390 0.9558 -0.0511 0.9558 0.9777
No log 11.8788 392 0.9295 0.0392 0.9295 0.9641
No log 11.9394 394 0.8600 0.1047 0.8600 0.9274
No log 12.0 396 0.8206 -0.0644 0.8206 0.9059
No log 12.0606 398 0.8216 -0.0976 0.8216 0.9064
No log 12.1212 400 0.8220 -0.1033 0.8220 0.9067
No log 12.1818 402 0.8373 -0.1230 0.8373 0.9151
No log 12.2424 404 0.8515 -0.1690 0.8515 0.9228
No log 12.3030 406 0.8859 -0.1951 0.8859 0.9412
No log 12.3636 408 0.9052 -0.1939 0.9052 0.9514
No log 12.4242 410 0.9011 -0.1527 0.9011 0.9493
No log 12.4848 412 0.9099 -0.0672 0.9099 0.9539
No log 12.5455 414 0.9653 -0.0408 0.9653 0.9825
No log 12.6061 416 0.9105 -0.0408 0.9105 0.9542
No log 12.6667 418 0.8494 0.0159 0.8494 0.9216
No log 12.7273 420 0.8137 0.0374 0.8137 0.9021
No log 12.7879 422 0.7991 0.0460 0.7991 0.8939
No log 12.8485 424 0.8038 -0.0329 0.8038 0.8965
No log 12.9091 426 0.7769 0.1021 0.7769 0.8814
No log 12.9697 428 0.7793 0.0071 0.7793 0.8828
No log 13.0303 430 0.7647 0.1565 0.7647 0.8744
No log 13.0909 432 0.7774 0.0031 0.7774 0.8817
No log 13.1515 434 0.8218 -0.0692 0.8218 0.9065
No log 13.2121 436 0.8519 -0.1843 0.8519 0.9230
No log 13.2727 438 0.7781 -0.0033 0.7781 0.8821
No log 13.3333 440 0.8169 0.0476 0.8169 0.9039
No log 13.3939 442 0.8316 0.0476 0.8316 0.9119
No log 13.4545 444 0.8475 -0.0426 0.8475 0.9206
No log 13.5152 446 0.9631 0.0433 0.9631 0.9814
No log 13.5758 448 0.9217 -0.1033 0.9217 0.9600
No log 13.6364 450 0.8259 -0.0532 0.8259 0.9088
No log 13.6970 452 0.8547 0.0099 0.8547 0.9245
No log 13.7576 454 0.9024 -0.0033 0.9024 0.9499
No log 13.8182 456 0.8315 0.0099 0.8315 0.9119
No log 13.8788 458 0.8312 0.0099 0.8312 0.9117
No log 13.9394 460 0.8222 -0.0145 0.8222 0.9067
No log 14.0 462 0.8441 -0.1331 0.8441 0.9187
No log 14.0606 464 0.8791 -0.0896 0.8791 0.9376
No log 14.1212 466 0.8434 0.0027 0.8434 0.9184
No log 14.1818 468 0.8479 -0.0274 0.8479 0.9208
No log 14.2424 470 0.8884 -0.1271 0.8884 0.9425
No log 14.3030 472 0.9379 -0.0843 0.9379 0.9685
No log 14.3636 474 0.8822 -0.1833 0.8822 0.9393
No log 14.4242 476 0.8330 0.0 0.8330 0.9127
No log 14.4848 478 0.9019 -0.0723 0.9019 0.9497
No log 14.5455 480 1.0265 0.0157 1.0265 1.0132
No log 14.6061 482 1.0093 0.0067 1.0093 1.0047
No log 14.6667 484 0.8734 -0.0757 0.8734 0.9345
No log 14.7273 486 0.9111 -0.0887 0.9111 0.9545
No log 14.7879 488 1.0908 -0.0996 1.0908 1.0444
No log 14.8485 490 1.1286 -0.0577 1.1286 1.0624
No log 14.9091 492 1.0055 0.0147 1.0055 1.0027
No log 14.9697 494 0.8497 -0.1277 0.8497 0.9218
No log 15.0303 496 0.8358 0.0786 0.8358 0.9142
No log 15.0909 498 0.8527 -0.0122 0.8527 0.9234
0.3061 15.1515 500 0.8659 -0.0373 0.8659 0.9305
0.3061 15.2121 502 0.8723 -0.0373 0.8723 0.9340
0.3061 15.2727 504 0.8603 -0.0686 0.8603 0.9275
0.3061 15.3333 506 0.8686 -0.1060 0.8686 0.9320
0.3061 15.3939 508 0.8946 -0.1077 0.8946 0.9458
0.3061 15.4545 510 0.9555 0.0016 0.9555 0.9775
0.3061 15.5152 512 1.0025 -0.0376 1.0025 1.0012
0.3061 15.5758 514 1.0176 -0.0682 1.0176 1.0088
0.3061 15.6364 516 1.0411 -0.1102 1.0411 1.0203
0.3061 15.6970 518 1.1103 -0.1579 1.1103 1.0537
0.3061 15.7576 520 1.1205 -0.0961 1.1205 1.0586

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k13_task3_organization

Finetuned
(4019)
this model