ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k16_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9605
  • Qwk: 0.6269
  • Mse: 0.9605
  • Rmse: 0.9801

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0167 2 6.7127 0.0242 6.7127 2.5909
No log 0.0333 4 4.6911 0.0784 4.6911 2.1659
No log 0.05 6 2.8781 0.0988 2.8781 1.6965
No log 0.0667 8 2.1077 0.1985 2.1077 1.4518
No log 0.0833 10 1.7193 0.2182 1.7193 1.3112
No log 0.1 12 1.5373 0.1622 1.5373 1.2399
No log 0.1167 14 1.5798 0.2414 1.5798 1.2569
No log 0.1333 16 1.6194 0.2564 1.6194 1.2726
No log 0.15 18 1.7043 0.2353 1.7043 1.3055
No log 0.1667 20 2.1637 0.2568 2.1637 1.4710
No log 0.1833 22 2.5505 0.1975 2.5505 1.5970
No log 0.2 24 1.8686 0.2687 1.8686 1.3670
No log 0.2167 26 1.2311 0.5077 1.2311 1.1095
No log 0.2333 28 1.1771 0.5156 1.1771 1.0849
No log 0.25 30 1.3199 0.4091 1.3199 1.1489
No log 0.2667 32 1.5741 0.4029 1.5741 1.2546
No log 0.2833 34 1.4570 0.4203 1.4570 1.2070
No log 0.3 36 1.2483 0.5344 1.2483 1.1173
No log 0.3167 38 1.3216 0.4098 1.3216 1.1496
No log 0.3333 40 1.3438 0.4032 1.3438 1.1592
No log 0.35 42 1.2958 0.4062 1.2958 1.1383
No log 0.3667 44 1.5407 0.2623 1.5407 1.2413
No log 0.3833 46 1.9201 0.3111 1.9201 1.3857
No log 0.4 48 1.8060 0.3358 1.8060 1.3439
No log 0.4167 50 1.3959 0.4366 1.3959 1.1815
No log 0.4333 52 1.0344 0.6 1.0344 1.0171
No log 0.45 54 1.2738 0.4444 1.2738 1.1286
No log 0.4667 56 1.3160 0.4127 1.3160 1.1472
No log 0.4833 58 1.0396 0.5931 1.0396 1.0196
No log 0.5 60 1.0846 0.6174 1.0846 1.0414
No log 0.5167 62 1.1916 0.6093 1.1916 1.0916
No log 0.5333 64 1.1441 0.5986 1.1441 1.0696
No log 0.55 66 1.0656 0.6259 1.0656 1.0323
No log 0.5667 68 1.0615 0.6174 1.0615 1.0303
No log 0.5833 70 1.1670 0.5772 1.1670 1.0803
No log 0.6 72 1.0649 0.5674 1.0649 1.0320
No log 0.6167 74 1.0437 0.5775 1.0437 1.0216
No log 0.6333 76 1.0560 0.6043 1.0560 1.0276
No log 0.65 78 0.9927 0.6043 0.9927 0.9964
No log 0.6667 80 0.9332 0.6286 0.9332 0.9660
No log 0.6833 82 0.9593 0.6619 0.9593 0.9795
No log 0.7 84 0.9306 0.6475 0.9306 0.9647
No log 0.7167 86 0.7948 0.7483 0.7948 0.8915
No log 0.7333 88 0.7513 0.7550 0.7513 0.8668
No log 0.75 90 0.7521 0.7226 0.7521 0.8672
No log 0.7667 92 0.7821 0.7089 0.7821 0.8844
No log 0.7833 94 0.7490 0.7143 0.7490 0.8654
No log 0.8 96 0.7724 0.7097 0.7724 0.8789
No log 0.8167 98 0.8228 0.6883 0.8228 0.9071
No log 0.8333 100 0.9002 0.6667 0.9002 0.9488
No log 0.85 102 0.8730 0.7248 0.8730 0.9344
No log 0.8667 104 0.9345 0.6809 0.9345 0.9667
No log 0.8833 106 0.8739 0.6933 0.8739 0.9348
No log 0.9 108 1.0123 0.6323 1.0123 1.0061
No log 0.9167 110 1.0190 0.6323 1.0190 1.0095
No log 0.9333 112 0.9186 0.6621 0.9186 0.9584
No log 0.95 114 0.9586 0.6154 0.9586 0.9791
No log 0.9667 116 0.9837 0.6043 0.9837 0.9918
No log 0.9833 118 0.9878 0.6294 0.9878 0.9939
No log 1.0 120 0.9813 0.6483 0.9813 0.9906
No log 1.0167 122 0.9212 0.6711 0.9212 0.9598
No log 1.0333 124 0.9914 0.675 0.9914 0.9957
No log 1.05 126 1.1074 0.6289 1.1074 1.0523
No log 1.0667 128 0.9811 0.6345 0.9811 0.9905
No log 1.0833 130 0.8816 0.6483 0.8816 0.9390
No log 1.1 132 0.7849 0.6621 0.7849 0.8859
No log 1.1167 134 0.7560 0.6849 0.7560 0.8695
No log 1.1333 136 0.7379 0.6892 0.7379 0.8590
No log 1.15 138 0.7244 0.6892 0.7244 0.8511
No log 1.1667 140 0.7396 0.6757 0.7396 0.8600
No log 1.1833 142 0.8064 0.7075 0.8064 0.8980
No log 1.2 144 0.6913 0.7368 0.6913 0.8314
No log 1.2167 146 0.7824 0.7117 0.7824 0.8846
No log 1.2333 148 0.8590 0.6914 0.8590 0.9268
No log 1.25 150 0.7000 0.7413 0.7000 0.8367
No log 1.2667 152 0.7310 0.7092 0.7310 0.8550
No log 1.2833 154 0.8307 0.6806 0.8307 0.9114
No log 1.3 156 0.8285 0.7114 0.8285 0.9102
No log 1.3167 158 0.8488 0.7237 0.8488 0.9213
No log 1.3333 160 0.8617 0.6759 0.8617 0.9283
No log 1.35 162 1.0389 0.5915 1.0389 1.0193
No log 1.3667 164 1.0747 0.5625 1.0747 1.0367
No log 1.3833 166 0.9042 0.6528 0.9042 0.9509
No log 1.4 168 0.8426 0.64 0.8426 0.9179
No log 1.4167 170 0.8019 0.7089 0.8019 0.8955
No log 1.4333 172 0.7810 0.6536 0.7810 0.8837
No log 1.45 174 0.8449 0.6748 0.8449 0.9192
No log 1.4667 176 0.7481 0.7097 0.7481 0.8649
No log 1.4833 178 0.7723 0.7034 0.7723 0.8788
No log 1.5 180 0.7367 0.7568 0.7367 0.8583
No log 1.5167 182 0.7072 0.7848 0.7072 0.8410
No log 1.5333 184 0.6630 0.7901 0.6630 0.8142
No log 1.55 186 0.6078 0.7857 0.6078 0.7796
No log 1.5667 188 0.5738 0.8114 0.5738 0.7575
No log 1.5833 190 0.5792 0.8193 0.5792 0.7610
No log 1.6 192 0.6306 0.7922 0.6306 0.7941
No log 1.6167 194 0.7117 0.7260 0.7117 0.8436
No log 1.6333 196 0.7931 0.7 0.7931 0.8905
No log 1.65 198 0.8079 0.7101 0.8079 0.8988
No log 1.6667 200 0.7869 0.7353 0.7869 0.8871
No log 1.6833 202 0.8728 0.6618 0.8728 0.9342
No log 1.7 204 0.8739 0.6277 0.8739 0.9348
No log 1.7167 206 0.7641 0.7246 0.7641 0.8741
No log 1.7333 208 0.7556 0.7338 0.7556 0.8692
No log 1.75 210 0.8494 0.6443 0.8494 0.9216
No log 1.7667 212 0.9436 0.6194 0.9436 0.9714
No log 1.7833 214 0.9031 0.6389 0.9031 0.9503
No log 1.8 216 0.8662 0.6519 0.8662 0.9307
No log 1.8167 218 0.8549 0.6667 0.8549 0.9246
No log 1.8333 220 0.8093 0.6912 0.8093 0.8996
No log 1.85 222 0.7302 0.7246 0.7302 0.8545
No log 1.8667 224 0.7070 0.7383 0.7070 0.8408
No log 1.8833 226 0.7265 0.7219 0.7265 0.8524
No log 1.9 228 0.6845 0.7821 0.6845 0.8273
No log 1.9167 230 0.6742 0.7486 0.6742 0.8211
No log 1.9333 232 0.6789 0.7314 0.6789 0.8240
No log 1.95 234 0.7016 0.7262 0.7016 0.8376
No log 1.9667 236 0.7865 0.7381 0.7865 0.8868
No log 1.9833 238 0.8914 0.6905 0.8914 0.9442
No log 2.0 240 0.8295 0.6753 0.8295 0.9108
No log 2.0167 242 0.8125 0.6622 0.8125 0.9014
No log 2.0333 244 0.7864 0.7222 0.7864 0.8868
No log 2.05 246 0.8253 0.7007 0.8253 0.9085
No log 2.0667 248 0.8410 0.6715 0.8410 0.9170
No log 2.0833 250 0.8408 0.6857 0.8408 0.9169
No log 2.1 252 0.8047 0.6667 0.8047 0.8970
No log 2.1167 254 0.7708 0.6887 0.7708 0.8780
No log 2.1333 256 0.7664 0.7305 0.7664 0.8754
No log 2.15 258 0.7858 0.7586 0.7858 0.8864
No log 2.1667 260 0.7795 0.7586 0.7795 0.8829
No log 2.1833 262 0.7379 0.7093 0.7379 0.8590
No log 2.2 264 0.7243 0.7195 0.7243 0.8510
No log 2.2167 266 0.7555 0.7179 0.7555 0.8692
No log 2.2333 268 0.7832 0.7 0.7832 0.8850
No log 2.25 270 0.7683 0.7299 0.7683 0.8765
No log 2.2667 272 0.7401 0.7092 0.7401 0.8603
No log 2.2833 274 0.7275 0.7 0.7275 0.8530
No log 2.3 276 0.6965 0.7671 0.6965 0.8346
No log 2.3167 278 0.7235 0.7333 0.7235 0.8506
No log 2.3333 280 0.7668 0.7170 0.7668 0.8757
No log 2.35 282 0.7544 0.72 0.7544 0.8685
No log 2.3667 284 0.7895 0.6906 0.7895 0.8885
No log 2.3833 286 0.8580 0.7 0.8580 0.9263
No log 2.4 288 0.9081 0.6519 0.9081 0.9530
No log 2.4167 290 1.0365 0.5507 1.0365 1.0181
No log 2.4333 292 1.0924 0.5517 1.0924 1.0452
No log 2.45 294 0.9800 0.5821 0.9800 0.9900
No log 2.4667 296 0.9004 0.6232 0.9004 0.9489
No log 2.4833 298 0.7898 0.6912 0.7898 0.8887
No log 2.5 300 0.8005 0.7234 0.8005 0.8947
No log 2.5167 302 0.8140 0.6897 0.8140 0.9022
No log 2.5333 304 0.7104 0.7432 0.7104 0.8428
No log 2.55 306 0.6295 0.7792 0.6295 0.7934
No log 2.5667 308 0.6730 0.7673 0.6730 0.8204
No log 2.5833 310 0.6809 0.7582 0.6809 0.8252
No log 2.6 312 0.6678 0.7586 0.6678 0.8172
No log 2.6167 314 0.7297 0.7 0.7297 0.8542
No log 2.6333 316 0.7407 0.7 0.7407 0.8607
No log 2.65 318 0.6889 0.7286 0.6889 0.8300
No log 2.6667 320 0.6978 0.7453 0.6978 0.8354
No log 2.6833 322 0.7029 0.7692 0.7029 0.8384
No log 2.7 324 0.7336 0.7586 0.7336 0.8565
No log 2.7167 326 0.7128 0.7456 0.7128 0.8443
No log 2.7333 328 0.7191 0.7190 0.7191 0.8480
No log 2.75 330 0.7423 0.7234 0.7423 0.8616
No log 2.7667 332 0.7825 0.6912 0.7825 0.8846
No log 2.7833 334 0.7954 0.6912 0.7954 0.8919
No log 2.8 336 0.7670 0.6901 0.7670 0.8758
No log 2.8167 338 0.7385 0.7067 0.7385 0.8594
No log 2.8333 340 0.8249 0.6988 0.8249 0.9082
No log 2.85 342 0.9728 0.7093 0.9728 0.9863
No log 2.8667 344 0.8879 0.6792 0.8879 0.9423
No log 2.8833 346 0.7459 0.72 0.7459 0.8636
No log 2.9 348 0.7127 0.7376 0.7127 0.8442
No log 2.9167 350 0.7158 0.7376 0.7158 0.8460
No log 2.9333 352 0.6992 0.7755 0.6992 0.8362
No log 2.95 354 0.7076 0.75 0.7076 0.8412
No log 2.9667 356 0.7423 0.7114 0.7423 0.8616
No log 2.9833 358 0.7294 0.7162 0.7294 0.8540
No log 3.0 360 0.7132 0.7234 0.7132 0.8445
No log 3.0167 362 0.6941 0.7660 0.6941 0.8331
No log 3.0333 364 0.6930 0.7571 0.6930 0.8325
No log 3.05 366 0.7316 0.7183 0.7316 0.8553
No log 3.0667 368 0.7497 0.6714 0.7497 0.8658
No log 3.0833 370 0.6878 0.7518 0.6878 0.8293
No log 3.1 372 0.6662 0.7692 0.6662 0.8162
No log 3.1167 374 0.6685 0.7347 0.6685 0.8176
No log 3.1333 376 0.8222 0.7152 0.8222 0.9067
No log 3.15 378 0.9716 0.6471 0.9716 0.9857
No log 3.1667 380 0.9377 0.6503 0.9377 0.9684
No log 3.1833 382 0.7636 0.7034 0.7636 0.8738
No log 3.2 384 0.6400 0.7724 0.6400 0.8000
No log 3.2167 386 0.6294 0.7552 0.6294 0.7933
No log 3.2333 388 0.6143 0.7552 0.6143 0.7838
No log 3.25 390 0.5931 0.8026 0.5931 0.7702
No log 3.2667 392 0.6143 0.7843 0.6143 0.7838
No log 3.2833 394 0.6368 0.8121 0.6368 0.7980
No log 3.3 396 0.6363 0.7974 0.6363 0.7977
No log 3.3167 398 0.6226 0.7945 0.6226 0.7891
No log 3.3333 400 0.6040 0.7832 0.6040 0.7772
No log 3.35 402 0.6050 0.8026 0.6050 0.7778
No log 3.3667 404 0.6139 0.8026 0.6139 0.7835
No log 3.3833 406 0.5980 0.8105 0.5980 0.7733
No log 3.4 408 0.6345 0.7778 0.6345 0.7966
No log 3.4167 410 0.7440 0.6986 0.7440 0.8626
No log 3.4333 412 0.7716 0.6986 0.7716 0.8784
No log 3.45 414 0.6797 0.7042 0.6797 0.8244
No log 3.4667 416 0.6523 0.7746 0.6523 0.8077
No log 3.4833 418 0.7191 0.7101 0.7191 0.8480
No log 3.5 420 0.7034 0.7286 0.7034 0.8387
No log 3.5167 422 0.6716 0.7801 0.6716 0.8195
No log 3.5333 424 0.7943 0.6950 0.7943 0.8912
No log 3.55 426 0.9862 0.6065 0.9862 0.9931
No log 3.5667 428 1.1073 0.6341 1.1073 1.0523
No log 3.5833 430 1.0035 0.6467 1.0035 1.0018
No log 3.6 432 0.7788 0.7027 0.7788 0.8825
No log 3.6167 434 0.6812 0.7534 0.6812 0.8254
No log 3.6333 436 0.6807 0.7448 0.6807 0.8250
No log 3.65 438 0.6852 0.7361 0.6852 0.8278
No log 3.6667 440 0.6809 0.7586 0.6809 0.8252
No log 3.6833 442 0.6879 0.7448 0.6879 0.8294
No log 3.7 444 0.7106 0.7467 0.7106 0.8430
No log 3.7167 446 0.7437 0.7211 0.7437 0.8624
No log 3.7333 448 0.7770 0.6892 0.7770 0.8815
No log 3.75 450 0.8665 0.6842 0.8665 0.9309
No log 3.7667 452 0.8899 0.6968 0.8899 0.9433
No log 3.7833 454 0.7754 0.7044 0.7754 0.8806
No log 3.8 456 0.7462 0.6800 0.7462 0.8639
No log 3.8167 458 0.7620 0.6944 0.7620 0.8729
No log 3.8333 460 0.7857 0.6944 0.7857 0.8864
No log 3.85 462 0.8087 0.7194 0.8087 0.8993
No log 3.8667 464 0.8088 0.7194 0.8088 0.8994
No log 3.8833 466 0.7804 0.7429 0.7804 0.8834
No log 3.9 468 0.7633 0.6809 0.7633 0.8737
No log 3.9167 470 0.7301 0.6853 0.7301 0.8545
No log 3.9333 472 0.7192 0.6846 0.7192 0.8480
No log 3.95 474 0.7287 0.6846 0.7287 0.8536
No log 3.9667 476 0.7396 0.6803 0.7396 0.8600
No log 3.9833 478 0.8120 0.6573 0.8120 0.9011
No log 4.0 480 0.9769 0.6443 0.9769 0.9884
No log 4.0167 482 1.0773 0.6174 1.0773 1.0379
No log 4.0333 484 0.9926 0.6143 0.9926 0.9963
No log 4.05 486 0.8364 0.6866 0.8364 0.9146
No log 4.0667 488 0.7772 0.6716 0.7772 0.8816
No log 4.0833 490 0.7422 0.6765 0.7422 0.8615
No log 4.1 492 0.7170 0.6765 0.7170 0.8468
No log 4.1167 494 0.8177 0.6974 0.8177 0.9043
No log 4.1333 496 0.8652 0.6709 0.8652 0.9302
No log 4.15 498 0.8480 0.6241 0.8480 0.9209
0.4142 4.1667 500 0.8179 0.6617 0.8179 0.9044
0.4142 4.1833 502 0.8454 0.6617 0.8454 0.9194
0.4142 4.2 504 0.8690 0.6364 0.8690 0.9322
0.4142 4.2167 506 0.8745 0.6364 0.8745 0.9351
0.4142 4.2333 508 0.8914 0.6364 0.8914 0.9442
0.4142 4.25 510 0.9605 0.6269 0.9605 0.9801

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k16_task1_organization

Finetuned
(4023)
this model