ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k16_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9437
  • Qwk: -0.1211
  • Mse: 0.9437
  • Rmse: 0.9715

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 3.5744 -0.0247 3.5744 1.8906
No log 0.1 4 1.7428 0.0772 1.7428 1.3202
No log 0.15 6 1.2865 0.0 1.2865 1.1343
No log 0.2 8 0.9801 -0.0385 0.9801 0.9900
No log 0.25 10 0.9432 -0.0013 0.9432 0.9712
No log 0.3 12 0.9716 -0.0301 0.9716 0.9857
No log 0.35 14 1.0805 -0.0012 1.0805 1.0395
No log 0.4 16 1.2473 -0.0241 1.2473 1.1168
No log 0.45 18 1.2641 -0.0234 1.2641 1.1243
No log 0.5 20 1.1927 0.0048 1.1927 1.0921
No log 0.55 22 1.0495 0.0196 1.0495 1.0245
No log 0.6 24 0.8781 0.0609 0.8781 0.9371
No log 0.65 26 0.8649 0.0260 0.8649 0.9300
No log 0.7 28 0.9597 -0.0117 0.9597 0.9796
No log 0.75 30 1.1594 -0.0221 1.1594 1.0767
No log 0.8 32 1.0936 -0.0234 1.0936 1.0458
No log 0.85 34 0.9065 -0.0301 0.9065 0.9521
No log 0.9 36 0.7650 -0.0662 0.7650 0.8746
No log 0.95 38 0.7201 0.0 0.7201 0.8486
No log 1.0 40 0.6737 0.0 0.6737 0.8208
No log 1.05 42 0.7805 0.0909 0.7805 0.8835
No log 1.1 44 1.6287 -0.0468 1.6287 1.2762
No log 1.15 46 1.7145 0.0016 1.7145 1.3094
No log 1.2 48 1.0334 -0.1283 1.0334 1.0166
No log 1.25 50 0.7481 0.1021 0.7481 0.8650
No log 1.3 52 0.7786 -0.0160 0.7786 0.8824
No log 1.35 54 0.7718 -0.1230 0.7718 0.8785
No log 1.4 56 0.7633 -0.1774 0.7633 0.8737
No log 1.45 58 0.7455 0.0 0.7455 0.8634
No log 1.5 60 0.7336 -0.0499 0.7336 0.8565
No log 1.55 62 0.7670 -0.0499 0.7670 0.8758
No log 1.6 64 0.9350 -0.1636 0.9350 0.9669
No log 1.65 66 0.9695 -0.0628 0.9695 0.9846
No log 1.7 68 0.8716 -0.0886 0.8716 0.9336
No log 1.75 70 0.8317 -0.0790 0.8317 0.9120
No log 1.8 72 0.8409 -0.1251 0.8409 0.9170
No log 1.85 74 0.8695 -0.0474 0.8695 0.9325
No log 1.9 76 0.8868 0.0676 0.8868 0.9417
No log 1.95 78 1.1033 -0.0658 1.1033 1.0504
No log 2.0 80 1.1922 -0.0385 1.1922 1.0919
No log 2.05 82 1.0649 -0.1238 1.0649 1.0319
No log 2.1 84 0.9259 0.0277 0.9259 0.9622
No log 2.15 86 0.9421 -0.0424 0.9421 0.9706
No log 2.2 88 1.0032 0.0545 1.0032 1.0016
No log 2.25 90 0.9030 0.0871 0.9030 0.9503
No log 2.3 92 0.7777 -0.1100 0.7777 0.8819
No log 2.35 94 0.7713 -0.1040 0.7713 0.8782
No log 2.4 96 0.8659 0.0871 0.8659 0.9305
No log 2.45 98 1.1417 -0.0539 1.1417 1.0685
No log 2.5 100 0.9678 -0.1257 0.9678 0.9838
No log 2.55 102 0.7695 -0.0091 0.7695 0.8772
No log 2.6 104 0.7827 0.0033 0.7827 0.8847
No log 2.65 106 0.8505 -0.1066 0.8505 0.9222
No log 2.7 108 1.0365 -0.1214 1.0365 1.0181
No log 2.75 110 0.9230 -0.0790 0.9230 0.9607
No log 2.8 112 0.7883 0.0318 0.7883 0.8879
No log 2.85 114 0.7733 0.1553 0.7733 0.8794
No log 2.9 116 0.8098 0.0956 0.8098 0.8999
No log 2.95 118 0.7505 -0.0062 0.7505 0.8663
No log 3.0 120 0.7936 0.0 0.7936 0.8908
No log 3.05 122 0.8609 0.0 0.8609 0.9279
No log 3.1 124 0.9310 -0.0322 0.9310 0.9649
No log 3.15 126 0.8571 0.0355 0.8571 0.9258
No log 3.2 128 0.8365 0.0209 0.8365 0.9146
No log 3.25 130 0.8122 -0.0406 0.8122 0.9012
No log 3.3 132 0.9313 -0.0456 0.9313 0.9650
No log 3.35 134 0.8412 0.0296 0.8412 0.9172
No log 3.4 136 0.7693 -0.0532 0.7693 0.8771
No log 3.45 138 0.7334 0.1021 0.7334 0.8564
No log 3.5 140 0.7370 0.0416 0.7370 0.8585
No log 3.55 142 0.7797 0.0334 0.7797 0.8830
No log 3.6 144 0.7471 0.1082 0.7471 0.8643
No log 3.65 146 0.7416 0.1082 0.7416 0.8612
No log 3.7 148 0.7652 0.1021 0.7652 0.8748
No log 3.75 150 0.7866 -0.0032 0.7866 0.8869
No log 3.8 152 0.8035 -0.0612 0.8035 0.8964
No log 3.85 154 0.8160 -0.1470 0.8160 0.9033
No log 3.9 156 0.8306 -0.1330 0.8306 0.9114
No log 3.95 158 0.8207 -0.0125 0.8207 0.9060
No log 4.0 160 0.8237 -0.0578 0.8237 0.9076
No log 4.05 162 0.8599 -0.0334 0.8599 0.9273
No log 4.1 164 0.8896 -0.0647 0.8896 0.9432
No log 4.15 166 1.0643 -0.0077 1.0643 1.0317
No log 4.2 168 1.0135 -0.0204 1.0135 1.0067
No log 4.25 170 0.9710 -0.1741 0.9710 0.9854
No log 4.3 172 0.9691 -0.0180 0.9691 0.9844
No log 4.35 174 1.0254 -0.0159 1.0254 1.0126
No log 4.4 176 0.9176 -0.1091 0.9176 0.9579
No log 4.45 178 0.8530 -0.0506 0.8530 0.9236
No log 4.5 180 0.8473 0.0432 0.8473 0.9205
No log 4.55 182 0.9410 -0.1194 0.9410 0.9701
No log 4.6 184 0.9467 -0.1194 0.9467 0.9730
No log 4.65 186 0.9260 -0.1187 0.9260 0.9623
No log 4.7 188 0.9856 -0.0336 0.9856 0.9928
No log 4.75 190 0.8996 -0.1176 0.8996 0.9485
No log 4.8 192 0.9705 -0.0331 0.9705 0.9852
No log 4.85 194 0.8886 -0.0264 0.8886 0.9426
No log 4.9 196 0.8623 -0.1204 0.8623 0.9286
No log 4.95 198 0.8549 -0.0483 0.8549 0.9246
No log 5.0 200 0.8678 -0.0628 0.8678 0.9316
No log 5.05 202 0.8979 -0.0533 0.8979 0.9476
No log 5.1 204 0.9535 -0.1874 0.9535 0.9765
No log 5.15 206 0.9136 -0.1393 0.9136 0.9558
No log 5.2 208 0.8553 -0.1969 0.8553 0.9248
No log 5.25 210 0.8158 -0.1527 0.8158 0.9032
No log 5.3 212 0.7664 0.0 0.7664 0.8754
No log 5.35 214 0.7900 -0.0179 0.7900 0.8888
No log 5.4 216 0.8596 0.0282 0.8596 0.9271
No log 5.45 218 0.8144 -0.0660 0.8144 0.9024
No log 5.5 220 0.8175 -0.0179 0.8175 0.9042
No log 5.55 222 0.8249 -0.1106 0.8249 0.9082
No log 5.6 224 0.9064 -0.0690 0.9064 0.9520
No log 5.65 226 0.9010 -0.1116 0.9010 0.9492
No log 5.7 228 0.9263 -0.1299 0.9263 0.9625
No log 5.75 230 0.9376 -0.1245 0.9376 0.9683
No log 5.8 232 0.9322 -0.1871 0.9322 0.9655
No log 5.85 234 1.2006 -0.0757 1.2006 1.0957
No log 5.9 236 1.5093 -0.1383 1.5093 1.2285
No log 5.95 238 1.2547 -0.0513 1.2547 1.1201
No log 6.0 240 0.8832 -0.0739 0.8832 0.9398
No log 6.05 242 0.8127 -0.1547 0.8127 0.9015
No log 6.1 244 0.8069 -0.2006 0.8069 0.8983
No log 6.15 246 0.7767 0.0 0.7767 0.8813
No log 6.2 248 0.8701 0.0129 0.8701 0.9328
No log 6.25 250 1.0689 -0.1665 1.0689 1.0339
No log 6.3 252 0.9893 -0.0842 0.9893 0.9946
No log 6.35 254 0.7984 0.0436 0.7984 0.8935
No log 6.4 256 0.7776 -0.2006 0.7776 0.8818
No log 6.45 258 0.7839 -0.2006 0.7839 0.8854
No log 6.5 260 0.7508 0.0 0.7508 0.8665
No log 6.55 262 0.8326 0.0714 0.8326 0.9125
No log 6.6 264 0.9395 -0.1261 0.9395 0.9693
No log 6.65 266 0.9281 -0.1251 0.9281 0.9634
No log 6.7 268 0.8906 -0.2062 0.8906 0.9437
No log 6.75 270 0.8864 -0.2008 0.8864 0.9415
No log 6.8 272 0.8849 -0.0984 0.8849 0.9407
No log 6.85 274 0.8936 -0.1172 0.8936 0.9453
No log 6.9 276 0.9298 -0.0939 0.9298 0.9643
No log 6.95 278 0.9852 -0.0572 0.9852 0.9926
No log 7.0 280 0.9714 -0.0373 0.9714 0.9856
No log 7.05 282 0.9024 -0.0370 0.9024 0.9499
No log 7.1 284 0.8648 0.0209 0.8648 0.9299
No log 7.15 286 0.8479 0.0209 0.8479 0.9208
No log 7.2 288 0.8349 -0.0488 0.8349 0.9137
No log 7.25 290 0.8255 -0.0062 0.8255 0.9086
No log 7.3 292 0.8586 -0.1208 0.8586 0.9266
No log 7.35 294 0.8767 -0.0730 0.8767 0.9363
No log 7.4 296 0.8990 -0.0999 0.8990 0.9481
No log 7.45 298 0.9349 -0.1905 0.9349 0.9669
No log 7.5 300 0.9705 -0.0180 0.9705 0.9851
No log 7.55 302 0.9729 -0.1278 0.9729 0.9864
No log 7.6 304 0.9217 -0.1452 0.9217 0.9600
No log 7.65 306 0.8567 -0.0643 0.8567 0.9256
No log 7.7 308 0.8153 -0.0113 0.8153 0.9029
No log 7.75 310 0.7990 0.0432 0.7990 0.8939
No log 7.8 312 0.8078 0.0670 0.8078 0.8988
No log 7.85 314 0.8410 0.0512 0.8410 0.9171
No log 7.9 316 0.8510 0.0512 0.8510 0.9225
No log 7.95 318 0.7814 0.0225 0.7814 0.8839
No log 8.0 320 0.8032 -0.0350 0.8032 0.8962
No log 8.05 322 0.8575 0.0249 0.8575 0.9260
No log 8.1 324 0.8303 -0.0350 0.8303 0.9112
No log 8.15 326 0.8418 -0.0524 0.8418 0.9175
No log 8.2 328 0.9120 -0.0295 0.9120 0.9550
No log 8.25 330 0.9298 -0.1755 0.9298 0.9643
No log 8.3 332 0.9563 -0.0622 0.9563 0.9779
No log 8.35 334 1.0364 0.0007 1.0364 1.0180
No log 8.4 336 1.0529 -0.0204 1.0529 1.0261
No log 8.45 338 0.9354 -0.0133 0.9354 0.9671
No log 8.5 340 0.9141 -0.1072 0.9141 0.9561
No log 8.55 342 0.9372 -0.1623 0.9372 0.9681
No log 8.6 344 0.8925 0.0053 0.8925 0.9447
No log 8.65 346 0.8747 -0.0831 0.8747 0.9352
No log 8.7 348 0.8450 -0.1040 0.8450 0.9192
No log 8.75 350 0.8334 -0.0612 0.8334 0.9129
No log 8.8 352 0.8475 -0.1459 0.8475 0.9206
No log 8.85 354 0.8872 -0.1333 0.8872 0.9419
No log 8.9 356 0.9413 -0.1452 0.9413 0.9702
No log 8.95 358 0.9935 -0.2425 0.9935 0.9968
No log 9.0 360 0.9720 -0.2698 0.9720 0.9859
No log 9.05 362 0.9763 -0.0248 0.9763 0.9881
No log 9.1 364 0.9180 -0.0350 0.9180 0.9581
No log 9.15 366 0.8894 -0.0387 0.8894 0.9431
No log 9.2 368 0.8890 -0.1457 0.8890 0.9429
No log 9.25 370 0.9174 -0.0373 0.9174 0.9578
No log 9.3 372 0.9446 -0.0408 0.9446 0.9719
No log 9.35 374 0.9711 -0.1894 0.9711 0.9855
No log 9.4 376 0.9710 -0.1894 0.9710 0.9854
No log 9.45 378 0.9653 -0.1066 0.9653 0.9825
No log 9.5 380 0.9427 -0.1060 0.9427 0.9709
No log 9.55 382 0.9306 -0.2381 0.9306 0.9647
No log 9.6 384 0.9357 -0.2381 0.9357 0.9673
No log 9.65 386 0.9723 -0.1730 0.9723 0.9860
No log 9.7 388 1.0015 0.0347 1.0015 1.0008
No log 9.75 390 0.9993 0.0285 0.9993 0.9996
No log 9.8 392 0.9365 -0.1715 0.9365 0.9678
No log 9.85 394 0.9032 -0.0209 0.9032 0.9504
No log 9.9 396 0.8564 -0.0567 0.8564 0.9254
No log 9.95 398 0.8269 -0.0958 0.8269 0.9093
No log 10.0 400 0.8246 -0.0949 0.8246 0.9081
No log 10.05 402 0.8292 -0.0488 0.8292 0.9106
No log 10.1 404 0.9071 -0.1518 0.9071 0.9524
No log 10.15 406 0.9883 -0.2316 0.9883 0.9941
No log 10.2 408 1.0084 -0.1163 1.0084 1.0042
No log 10.25 410 1.0992 -0.0159 1.0992 1.0484
No log 10.3 412 1.1429 -0.0670 1.1429 1.0691
No log 10.35 414 1.0497 -0.0056 1.0497 1.0245
No log 10.4 416 0.9522 -0.1755 0.9522 0.9758
No log 10.45 418 0.8999 -0.2116 0.8999 0.9486
No log 10.5 420 0.8676 -0.1176 0.8676 0.9315
No log 10.55 422 0.8815 0.0085 0.8815 0.9389
No log 10.6 424 0.9280 0.0285 0.9280 0.9633
No log 10.65 426 0.9596 0.0347 0.9596 0.9796
No log 10.7 428 0.9504 0.0285 0.9504 0.9749
No log 10.75 430 0.9087 0.0192 0.9087 0.9532
No log 10.8 432 0.8922 -0.0334 0.8922 0.9446
No log 10.85 434 0.8640 -0.0567 0.8640 0.9295
No log 10.9 436 0.8378 -0.0550 0.8378 0.9153
No log 10.95 438 0.8355 -0.0387 0.8355 0.9141
No log 11.0 440 0.8540 -0.0837 0.8540 0.9241
No log 11.05 442 0.8507 -0.0837 0.8507 0.9224
No log 11.1 444 0.8493 -0.0152 0.8493 0.9216
No log 11.15 446 0.8879 -0.1249 0.8879 0.9423
No log 11.2 448 0.8909 -0.1249 0.8909 0.9439
No log 11.25 450 0.8285 0.0807 0.8285 0.9102
No log 11.3 452 0.8223 0.0436 0.8223 0.9068
No log 11.35 454 0.8250 -0.0967 0.8250 0.9083
No log 11.4 456 0.8163 0.0 0.8163 0.9035
No log 11.45 458 0.7920 0.0031 0.7920 0.8900
No log 11.5 460 0.7797 0.1021 0.7797 0.8830
No log 11.55 462 0.8622 -0.0287 0.8622 0.9286
No log 11.6 464 0.9138 -0.0823 0.9138 0.9559
No log 11.65 466 0.8647 0.0807 0.8647 0.9299
No log 11.7 468 0.8348 0.0030 0.8348 0.9137
No log 11.75 470 0.8665 0.0610 0.8665 0.9309
No log 11.8 472 0.8938 -0.0259 0.8938 0.9454
No log 11.85 474 0.8997 0.0157 0.8997 0.9486
No log 11.9 476 0.8602 -0.0550 0.8602 0.9275
No log 11.95 478 0.8104 0.0967 0.8104 0.9002
No log 12.0 480 0.7848 0.1023 0.7848 0.8859
No log 12.05 482 0.7718 0.1023 0.7718 0.8785
No log 12.1 484 0.7817 0.0909 0.7817 0.8841
No log 12.15 486 0.7999 0.0857 0.7999 0.8944
No log 12.2 488 0.7977 0.0909 0.7977 0.8932
No log 12.25 490 0.8346 0.0857 0.8346 0.9136
No log 12.3 492 0.8355 0.0909 0.8355 0.9141
No log 12.35 494 0.8539 0.0857 0.8539 0.9241
No log 12.4 496 0.8701 0.0863 0.8701 0.9328
No log 12.45 498 0.8943 0.0318 0.8943 0.9457
0.2959 12.5 500 0.8760 -0.0059 0.8760 0.9360
0.2959 12.55 502 0.8510 0.0 0.8510 0.9225
0.2959 12.6 504 0.8432 0.0610 0.8432 0.9183
0.2959 12.65 506 0.7970 -0.0473 0.7970 0.8928
0.2959 12.7 508 0.7729 0.0914 0.7729 0.8791
0.2959 12.75 510 0.7822 0.0909 0.7822 0.8844
0.2959 12.8 512 0.7840 0.0909 0.7840 0.8854
0.2959 12.85 514 0.7853 0.1023 0.7853 0.8861
0.2959 12.9 516 0.8160 -0.0428 0.8160 0.9033
0.2959 12.95 518 0.8139 0.0030 0.8139 0.9021
0.2959 13.0 520 0.8123 0.0914 0.8123 0.9013
0.2959 13.05 522 0.8449 0.0857 0.8449 0.9192
0.2959 13.1 524 0.8591 -0.0753 0.8591 0.9269
0.2959 13.15 526 0.8508 -0.0215 0.8508 0.9224
0.2959 13.2 528 0.8466 -0.0264 0.8466 0.9201
0.2959 13.25 530 0.8314 0.0260 0.8314 0.9118
0.2959 13.3 532 0.8430 -0.0086 0.8430 0.9182
0.2959 13.35 534 0.8766 -0.0059 0.8766 0.9363
0.2959 13.4 536 0.9495 -0.0543 0.9495 0.9744
0.2959 13.45 538 0.9785 -0.0138 0.9785 0.9892
0.2959 13.5 540 0.9437 -0.1211 0.9437 0.9715

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k16_task3_organization

Finetuned
(4019)
this model