ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k12_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7986
  • Qwk: -0.0560
  • Mse: 0.7986
  • Rmse: 0.8936

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0667 2 3.6294 0.0048 3.6294 1.9051
No log 0.1333 4 2.5155 -0.0156 2.5155 1.5860
No log 0.2 6 2.2231 0.0104 2.2231 1.4910
No log 0.2667 8 1.5380 -0.0265 1.5380 1.2402
No log 0.3333 10 1.0413 0.0046 1.0413 1.0204
No log 0.4 12 1.0127 -0.0079 1.0127 1.0064
No log 0.4667 14 1.2962 0.0204 1.2962 1.1385
No log 0.5333 16 1.6088 0.0074 1.6088 1.2684
No log 0.6 18 1.6363 0.0312 1.6363 1.2792
No log 0.6667 20 0.8302 -0.0351 0.8302 0.9112
No log 0.7333 22 0.7663 0.0670 0.7663 0.8754
No log 0.8 24 1.2517 -0.0164 1.2517 1.1188
No log 0.8667 26 1.3392 -0.0207 1.3392 1.1572
No log 0.9333 28 0.8658 0.0515 0.8658 0.9305
No log 1.0 30 0.7135 0.0460 0.7135 0.8447
No log 1.0667 32 0.7388 -0.0240 0.7388 0.8596
No log 1.1333 34 0.8527 -0.0842 0.8527 0.9234
No log 1.2 36 0.9719 0.0680 0.9719 0.9858
No log 1.2667 38 0.9447 -0.1270 0.9447 0.9719
No log 1.3333 40 0.8797 0.0017 0.8797 0.9379
No log 1.4 42 0.8288 -0.0331 0.8288 0.9104
No log 1.4667 44 0.8407 -0.0331 0.8407 0.9169
No log 1.5333 46 0.8964 -0.1263 0.8964 0.9468
No log 1.6 48 1.0909 -0.0943 1.0909 1.0444
No log 1.6667 50 1.0162 -0.0936 1.0162 1.0080
No log 1.7333 52 0.9093 -0.1263 0.9093 0.9536
No log 1.8 54 0.9833 0.0741 0.9833 0.9916
No log 1.8667 56 0.9131 0.1273 0.9131 0.9555
No log 1.9333 58 0.7442 -0.0725 0.7442 0.8627
No log 2.0 60 0.9019 -0.1259 0.9019 0.9497
No log 2.0667 62 1.5403 -0.0253 1.5403 1.2411
No log 2.1333 64 1.5824 -0.0247 1.5824 1.2579
No log 2.2 66 1.4761 0.0048 1.4761 1.2149
No log 2.2667 68 1.1569 -0.0435 1.1569 1.0756
No log 2.3333 70 0.9739 -0.1267 0.9739 0.9869
No log 2.4 72 1.2216 -0.0423 1.2216 1.1053
No log 2.4667 74 1.1615 -0.0987 1.1615 1.0777
No log 2.5333 76 1.1054 -0.0658 1.1054 1.0514
No log 2.6 78 0.8846 -0.0240 0.8846 0.9405
No log 2.6667 80 0.7834 -0.0032 0.7834 0.8851
No log 2.7333 82 0.9117 0.0146 0.9117 0.9548
No log 2.8 84 0.8867 0.0175 0.8867 0.9416
No log 2.8667 86 0.8497 -0.1040 0.8497 0.9218
No log 2.9333 88 1.0399 -0.0380 1.0399 1.0198
No log 3.0 90 1.0237 -0.0378 1.0237 1.0118
No log 3.0667 92 0.8732 0.0749 0.8732 0.9345
No log 3.1333 94 0.8726 0.0277 0.8726 0.9341
No log 3.2 96 0.8295 0.0428 0.8295 0.9108
No log 3.2667 98 0.7897 0.1028 0.7897 0.8887
No log 3.3333 100 0.8488 0.0562 0.8488 0.9213
No log 3.4 102 0.8152 0.0562 0.8152 0.9029
No log 3.4667 104 0.8117 0.0600 0.8117 0.9009
No log 3.5333 106 0.7943 0.0610 0.7943 0.8912
No log 3.6 108 0.8544 0.0229 0.8544 0.9243
No log 3.6667 110 0.8290 -0.0029 0.8290 0.9105
No log 3.7333 112 0.9843 -0.0408 0.9843 0.9921
No log 3.8 114 0.8293 -0.0170 0.8293 0.9107
No log 3.8667 116 0.7872 -0.1010 0.7872 0.8872
No log 3.9333 118 0.7921 -0.1067 0.7921 0.8900
No log 4.0 120 0.7724 -0.0627 0.7724 0.8789
No log 4.0667 122 0.9297 0.0470 0.9297 0.9642
No log 4.1333 124 0.8456 -0.1094 0.8456 0.9195
No log 4.2 126 0.8180 -0.0493 0.8180 0.9044
No log 4.2667 128 0.8320 -0.1094 0.8320 0.9121
No log 4.3333 130 0.8970 -0.1106 0.8970 0.9471
No log 4.4 132 0.8592 -0.0612 0.8592 0.9270
No log 4.4667 134 0.8812 -0.0675 0.8812 0.9387
No log 4.5333 136 0.8211 -0.1026 0.8211 0.9062
No log 4.6 138 0.8255 -0.0493 0.8255 0.9085
No log 4.6667 140 0.8166 -0.1033 0.8166 0.9036
No log 4.7333 142 0.8262 -0.1100 0.8262 0.9089
No log 4.8 144 0.8538 -0.0612 0.8538 0.9240
No log 4.8667 146 0.8531 -0.0762 0.8531 0.9236
No log 4.9333 148 0.8686 -0.0750 0.8686 0.9320
No log 5.0 150 0.8253 -0.1397 0.8253 0.9085
No log 5.0667 152 0.8579 0.0247 0.8579 0.9262
No log 5.1333 154 0.9408 -0.0316 0.9408 0.9700
No log 5.2 156 0.8464 -0.0939 0.8464 0.9200
No log 5.2667 158 0.8933 -0.0214 0.8933 0.9451
No log 5.3333 160 0.8973 -0.0150 0.8973 0.9473
No log 5.4 162 0.8085 -0.1397 0.8085 0.8992
No log 5.4667 164 0.8537 0.0639 0.8537 0.9240
No log 5.5333 166 0.8037 0.0814 0.8037 0.8965
No log 5.6 168 0.8503 -0.1091 0.8503 0.9221
No log 5.6667 170 0.9227 0.1877 0.9227 0.9606
No log 5.7333 172 0.8878 0.0392 0.8878 0.9422
No log 5.8 174 0.7995 -0.0958 0.7995 0.8942
No log 5.8667 176 0.9331 0.0016 0.9331 0.9660
No log 5.9333 178 0.9806 0.0277 0.9806 0.9903
No log 6.0 180 0.7831 0.1902 0.7831 0.8849
No log 6.0667 182 0.7421 -0.0520 0.7421 0.8614
No log 6.1333 184 0.7358 -0.0520 0.7358 0.8578
No log 6.2 186 0.7450 -0.0541 0.7450 0.8631
No log 6.2667 188 0.7889 -0.0541 0.7889 0.8882
No log 6.3333 190 0.9208 -0.0778 0.9208 0.9596
No log 6.4 192 0.9544 -0.1088 0.9544 0.9769
No log 6.4667 194 0.9927 -0.0322 0.9927 0.9963
No log 6.5333 196 1.2425 0.0059 1.2425 1.1147
No log 6.6 198 1.2780 0.0074 1.2780 1.1305
No log 6.6667 200 1.0837 -0.0096 1.0837 1.0410
No log 6.7333 202 0.8832 -0.1951 0.8832 0.9398
No log 6.8 204 0.8320 -0.1535 0.8320 0.9121
No log 6.8667 206 0.8180 0.0031 0.8180 0.9044
No log 6.9333 208 0.8123 -0.0363 0.8123 0.9013
No log 7.0 210 0.7828 -0.0363 0.7828 0.8848
No log 7.0667 212 0.7524 -0.0520 0.7524 0.8674
No log 7.1333 214 0.8019 -0.0188 0.8019 0.8955
No log 7.2 216 0.8172 -0.0188 0.8172 0.9040
No log 7.2667 218 0.7902 -0.0062 0.7902 0.8890
No log 7.3333 220 0.8899 -0.1142 0.8899 0.9434
No log 7.4 222 0.9800 0.0092 0.9800 0.9899
No log 7.4667 224 0.9379 -0.0492 0.9379 0.9685
No log 7.5333 226 0.8978 -0.1330 0.8978 0.9475
No log 7.6 228 0.8941 -0.1330 0.8941 0.9456
No log 7.6667 230 0.9029 -0.1671 0.9029 0.9502
No log 7.7333 232 0.9284 -0.0442 0.9284 0.9635
No log 7.8 234 0.9559 -0.0363 0.9559 0.9777
No log 7.8667 236 0.9190 -0.0581 0.9190 0.9586
No log 7.9333 238 0.8986 -0.1464 0.8986 0.9479
No log 8.0 240 0.8809 -0.1466 0.8809 0.9386
No log 8.0667 242 0.8710 -0.1464 0.8710 0.9333
No log 8.1333 244 0.8533 -0.0226 0.8533 0.9237
No log 8.2 246 0.8707 0.0724 0.8707 0.9331
No log 8.2667 248 0.8754 -0.0181 0.8754 0.9356
No log 8.3333 250 0.8897 0.0710 0.8897 0.9433
No log 8.4 252 0.8740 0.0240 0.8740 0.9349
No log 8.4667 254 0.8816 -0.0226 0.8816 0.9389
No log 8.5333 256 0.9267 -0.0063 0.9267 0.9626
No log 8.6 258 0.9863 0.0392 0.9863 0.9931
No log 8.6667 260 1.0152 0.0125 1.0152 1.0076
No log 8.7333 262 0.9955 -0.1016 0.9955 0.9978
No log 8.8 264 0.9998 -0.0228 0.9998 0.9999
No log 8.8667 266 1.0675 -0.0120 1.0675 1.0332
No log 8.9333 268 1.0325 -0.0464 1.0325 1.0161
No log 9.0 270 0.9646 -0.0781 0.9646 0.9821
No log 9.0667 272 0.9895 -0.1841 0.9895 0.9947
No log 9.1333 274 0.9719 -0.0536 0.9719 0.9859
No log 9.2 276 0.9437 -0.0393 0.9437 0.9714
No log 9.2667 278 0.9137 0.0734 0.9137 0.9559
No log 9.3333 280 0.8594 0.0240 0.8594 0.9270
No log 9.4 282 0.8650 -0.1397 0.8650 0.9300
No log 9.4667 284 1.0209 -0.0854 1.0209 1.0104
No log 9.5333 286 1.0731 -0.0943 1.0731 1.0359
No log 9.6 288 0.9681 -0.0532 0.9681 0.9839
No log 9.6667 290 0.8749 -0.1211 0.8749 0.9353
No log 9.7333 292 0.9668 0.0092 0.9668 0.9833
No log 9.8 294 1.0082 -0.0120 1.0082 1.0041
No log 9.8667 296 0.9701 -0.0236 0.9701 0.9849
No log 9.9333 298 0.8733 -0.1671 0.8733 0.9345
No log 10.0 300 0.8402 -0.1395 0.8402 0.9166
No log 10.0667 302 0.8255 -0.1463 0.8255 0.9085
No log 10.1333 304 0.8231 -0.1882 0.8231 0.9072
No log 10.2 306 0.8467 -0.0692 0.8467 0.9202
No log 10.2667 308 0.9156 0.0331 0.9156 0.9569
No log 10.3333 310 0.9654 -0.0279 0.9654 0.9825
No log 10.4 312 0.9094 -0.0977 0.9094 0.9536
No log 10.4667 314 0.9433 0.2194 0.9433 0.9712
No log 10.5333 316 0.9057 0.0608 0.9057 0.9517
No log 10.6 318 0.9329 -0.1471 0.9329 0.9659
No log 10.6667 320 1.0300 0.0205 1.0300 1.0149
No log 10.7333 322 0.9981 -0.0551 0.9981 0.9991
No log 10.8 324 0.8743 -0.1601 0.8743 0.9351
No log 10.8667 326 0.8160 -0.1398 0.8160 0.9033
No log 10.9333 328 0.7835 -0.0473 0.7835 0.8851
No log 11.0 330 0.7700 -0.0473 0.7700 0.8775
No log 11.0667 332 0.7664 -0.0473 0.7664 0.8754
No log 11.1333 334 0.7853 0.0571 0.7853 0.8862
No log 11.2 336 0.8187 -0.1398 0.8187 0.9048
No log 11.2667 338 0.8796 -0.1204 0.8796 0.9379
No log 11.3333 340 0.8952 -0.1091 0.8952 0.9461
No log 11.4 342 0.8718 -0.1601 0.8718 0.9337
No log 11.4667 344 0.8400 -0.1330 0.8400 0.9165
No log 11.5333 346 0.8307 -0.0912 0.8307 0.9114
No log 11.6 348 0.8410 -0.0912 0.8410 0.9170
No log 11.6667 350 0.8363 -0.1398 0.8363 0.9145
No log 11.7333 352 0.8923 0.0279 0.8923 0.9446
No log 11.8 354 0.8987 0.1123 0.8987 0.9480
No log 11.8667 356 0.8411 -0.1398 0.8411 0.9171
No log 11.9333 358 0.7904 -0.1398 0.7904 0.8891
No log 12.0 360 0.7704 -0.0578 0.7704 0.8777
No log 12.0667 362 0.8019 -0.0692 0.8019 0.8955
No log 12.1333 364 0.8852 0.0347 0.8852 0.9409
No log 12.2 366 0.9729 0.0219 0.9729 0.9863
No log 12.2667 368 1.0014 0.0516 1.0014 1.0007
No log 12.3333 370 1.0430 0.0534 1.0430 1.0213
No log 12.4 372 0.9610 0.0451 0.9610 0.9803
No log 12.4667 374 0.8942 -0.0517 0.8942 0.9456
No log 12.5333 376 0.8609 -0.1523 0.8609 0.9278
No log 12.6 378 0.8749 -0.0089 0.8749 0.9354
No log 12.6667 380 0.9320 0.0770 0.9320 0.9654
No log 12.7333 382 0.9066 0.0361 0.9066 0.9522
No log 12.8 384 0.8183 -0.1399 0.8183 0.9046
No log 12.8667 386 0.7290 0.0033 0.7290 0.8538
No log 12.9333 388 0.7224 0.0395 0.7224 0.8499
No log 13.0 390 0.7527 0.0334 0.7527 0.8676
No log 13.0667 392 0.7648 0.0964 0.7648 0.8745
No log 13.1333 394 0.8099 -0.1397 0.8099 0.9000
No log 13.2 396 0.8758 -0.0144 0.8758 0.9358
No log 13.2667 398 0.8723 -0.0204 0.8723 0.9340
No log 13.3333 400 0.8638 -0.1266 0.8638 0.9294
No log 13.4 402 0.8811 -0.1832 0.8811 0.9386
No log 13.4667 404 0.9400 -0.1013 0.9400 0.9695
No log 13.5333 406 0.9574 -0.1335 0.9574 0.9785
No log 13.6 408 0.9296 -0.0138 0.9296 0.9641
No log 13.6667 410 0.9781 0.0418 0.9781 0.9890
No log 13.7333 412 0.9786 0.0428 0.9786 0.9893
No log 13.8 414 0.9403 -0.0284 0.9403 0.9697
No log 13.8667 416 0.9074 -0.0784 0.9074 0.9526
No log 13.9333 418 0.8760 0.0 0.8760 0.9360
No log 14.0 420 0.8804 -0.1051 0.8804 0.9383
No log 14.0667 422 0.9131 0.0754 0.9131 0.9556
No log 14.1333 424 0.9079 -0.0393 0.9079 0.9528
No log 14.2 426 0.8627 -0.1270 0.8627 0.9288
No log 14.2667 428 0.8410 -0.1882 0.8410 0.9171
No log 14.3333 430 0.8261 -0.0595 0.8261 0.9089
No log 14.4 432 0.8294 -0.1594 0.8294 0.9107
No log 14.4667 434 0.8220 -0.1616 0.8220 0.9067
No log 14.5333 436 0.8505 -0.0786 0.8505 0.9222
No log 14.6 438 0.9228 -0.0477 0.9228 0.9606
No log 14.6667 440 1.0316 0.0164 1.0316 1.0157
No log 14.7333 442 1.0654 -0.1223 1.0654 1.0322
No log 14.8 444 1.0039 -0.0181 1.0039 1.0020
No log 14.8667 446 0.9026 -0.1659 0.9026 0.9500
No log 14.9333 448 0.9107 -0.1126 0.9107 0.9543
No log 15.0 450 0.8954 -0.1126 0.8954 0.9462
No log 15.0667 452 0.8233 -0.1616 0.8233 0.9074
No log 15.1333 454 0.8533 -0.1398 0.8533 0.9238
No log 15.2 456 0.9054 0.1123 0.9054 0.9515
No log 15.2667 458 0.9373 0.0377 0.9373 0.9681
No log 15.3333 460 0.9345 0.0007 0.9345 0.9667
No log 15.4 462 0.8976 -0.1410 0.8976 0.9474
No log 15.4667 464 0.9346 -0.0212 0.9346 0.9667
No log 15.5333 466 0.9673 -0.1079 0.9673 0.9835
No log 15.6 468 0.9254 -0.0923 0.9254 0.9620
No log 15.6667 470 0.9462 -0.0353 0.9462 0.9727
No log 15.7333 472 0.9561 0.0086 0.9561 0.9778
No log 15.8 474 0.8990 0.0229 0.8990 0.9482
No log 15.8667 476 0.8449 -0.0849 0.8449 0.9192
No log 15.9333 478 0.8162 -0.1081 0.8162 0.9034
No log 16.0 480 0.8159 -0.1018 0.8159 0.9033
No log 16.0667 482 0.8290 -0.0366 0.8290 0.9105
No log 16.1333 484 0.8302 0.0 0.8302 0.9112
No log 16.2 486 0.8164 0.0376 0.8164 0.9035
No log 16.2667 488 0.8015 -0.0578 0.8015 0.8952
No log 16.3333 490 0.7784 -0.1622 0.7784 0.8822
No log 16.4 492 0.7585 -0.1616 0.7585 0.8709
No log 16.4667 494 0.7563 0.0863 0.7563 0.8696
No log 16.5333 496 0.7476 -0.1094 0.7476 0.8646
No log 16.6 498 0.7462 -0.1018 0.7462 0.8639
0.2905 16.6667 500 0.7629 -0.0473 0.7629 0.8734
0.2905 16.7333 502 0.8002 -0.1470 0.8002 0.8946
0.2905 16.8 504 0.8472 -0.1470 0.8472 0.9204
0.2905 16.8667 506 0.8644 -0.1823 0.8644 0.9297
0.2905 16.9333 508 0.8208 -0.1470 0.8208 0.9060
0.2905 17.0 510 0.7986 -0.0560 0.7986 0.8936

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k12_task3_organization

Finetuned
(4019)
this model