ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k12_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9908
  • Qwk: -0.0076
  • Mse: 0.9908
  • Rmse: 0.9954

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0667 2 3.5131 -0.0047 3.5131 1.8743
No log 0.1333 4 2.4381 -0.0136 2.4381 1.5614
No log 0.2 6 1.8738 -0.0015 1.8738 1.3689
No log 0.2667 8 1.6415 0.0213 1.6415 1.2812
No log 0.3333 10 1.0148 0.0493 1.0148 1.0074
No log 0.4 12 0.7753 -0.1241 0.7753 0.8805
No log 0.4667 14 0.8139 -0.1241 0.8139 0.9021
No log 0.5333 16 0.8538 -0.1257 0.8538 0.9240
No log 0.6 18 1.1198 -0.0234 1.1198 1.0582
No log 0.6667 20 1.1956 -0.0234 1.1956 1.0934
No log 0.7333 22 1.1666 -0.0468 1.1666 1.0801
No log 0.8 24 1.3179 -0.0490 1.3179 1.1480
No log 0.8667 26 1.2304 -0.0479 1.2304 1.1092
No log 0.9333 28 1.0196 0.1120 1.0196 1.0097
No log 1.0 30 0.9136 -0.0852 0.9136 0.9558
No log 1.0667 32 0.9139 -0.1257 0.9139 0.9560
No log 1.1333 34 1.0957 -0.0117 1.0957 1.0468
No log 1.2 36 1.1909 -0.0728 1.1909 1.0913
No log 1.2667 38 1.2048 -0.0490 1.2048 1.0976
No log 1.3333 40 1.1212 -0.0207 1.1212 1.0589
No log 1.4 42 1.0794 -0.0207 1.0794 1.0389
No log 1.4667 44 1.0006 -0.0117 1.0006 1.0003
No log 1.5333 46 0.9367 -0.0031 0.9367 0.9678
No log 1.6 48 0.9499 -0.0982 0.9499 0.9746
No log 1.6667 50 0.9470 -0.0949 0.9470 0.9732
No log 1.7333 52 0.9841 -0.0972 0.9841 0.9920
No log 1.8 54 0.8885 -0.0870 0.8885 0.9426
No log 1.8667 56 0.8609 -0.0833 0.8609 0.9278
No log 1.9333 58 0.9141 -0.0474 0.9141 0.9561
No log 2.0 60 1.2422 0.0080 1.2422 1.1145
No log 2.0667 62 1.5007 -0.0247 1.5007 1.2250
No log 2.1333 64 0.9828 -0.0923 0.9828 0.9914
No log 2.2 66 0.9474 -0.1270 0.9474 0.9734
No log 2.2667 68 1.1467 -0.0149 1.1467 1.0708
No log 2.3333 70 1.2617 -0.0468 1.2617 1.1232
No log 2.4 72 1.4061 -0.0490 1.4061 1.1858
No log 2.4667 74 1.1255 -0.0648 1.1255 1.0609
No log 2.5333 76 0.9263 -0.0056 0.9263 0.9625
No log 2.6 78 0.8328 -0.1738 0.8328 0.9126
No log 2.6667 80 0.8765 -0.2201 0.8765 0.9362
No log 2.7333 82 1.2316 0.0107 1.2316 1.1098
No log 2.8 84 1.0828 -0.0097 1.0828 1.0406
No log 2.8667 86 0.7884 -0.1765 0.7884 0.8879
No log 2.9333 88 0.7640 -0.1765 0.7640 0.8740
No log 3.0 90 0.8560 -0.0371 0.8560 0.9252
No log 3.0667 92 0.9792 -0.0143 0.9792 0.9895
No log 3.1333 94 1.1217 -0.0236 1.1217 1.0591
No log 3.2 96 0.8629 0.0099 0.8629 0.9289
No log 3.2667 98 0.8021 0.0282 0.8021 0.8956
No log 3.3333 100 0.9563 0.0748 0.9563 0.9779
No log 3.4 102 1.5886 -0.0157 1.5886 1.2604
No log 3.4667 104 1.4921 -0.0367 1.4921 1.2215
No log 3.5333 106 0.8357 0.1047 0.8357 0.9142
No log 3.6 108 0.7572 -0.0069 0.7572 0.8702
No log 3.6667 110 0.7802 -0.1158 0.7802 0.8833
No log 3.7333 112 0.8736 0.0191 0.8736 0.9346
No log 3.8 114 0.9528 0.0099 0.9528 0.9761
No log 3.8667 116 0.8735 -0.0264 0.8735 0.9346
No log 3.9333 118 0.9288 0.0099 0.9288 0.9638
No log 4.0 120 0.9591 0.0071 0.9591 0.9793
No log 4.0667 122 0.9907 0.0476 0.9907 0.9953
No log 4.1333 124 1.0787 0.1064 1.0787 1.0386
No log 4.2 126 1.1054 0.0587 1.1054 1.0514
No log 4.2667 128 1.0263 -0.0076 1.0263 1.0131
No log 4.3333 130 0.8172 -0.0690 0.8172 0.9040
No log 4.4 132 0.8139 -0.0628 0.8139 0.9022
No log 4.4667 134 0.8761 0.0956 0.8761 0.9360
No log 4.5333 136 0.9808 0.0651 0.9808 0.9904
No log 4.6 138 1.1032 -0.0925 1.1032 1.0503
No log 4.6667 140 0.9904 0.0200 0.9904 0.9952
No log 4.7333 142 0.8214 0.1495 0.8214 0.9063
No log 4.8 144 0.8226 0.1001 0.8226 0.9070
No log 4.8667 146 0.7704 0.0863 0.7704 0.8777
No log 4.9333 148 0.8790 0.0525 0.8790 0.9375
No log 5.0 150 1.0296 0.0304 1.0296 1.0147
No log 5.0667 152 0.8162 -0.0228 0.8162 0.9034
No log 5.1333 154 0.7993 -0.1081 0.7993 0.8940
No log 5.2 156 0.8154 -0.1100 0.8154 0.9030
No log 5.2667 158 0.9980 0.0424 0.9980 0.9990
No log 5.3333 160 1.1567 -0.0877 1.1567 1.0755
No log 5.4 162 1.2509 -0.1848 1.2509 1.1184
No log 5.4667 164 1.0405 -0.0054 1.0405 1.0200
No log 5.5333 166 0.8549 -0.1535 0.8549 0.9246
No log 5.6 168 0.8914 -0.1204 0.8914 0.9442
No log 5.6667 170 0.8544 -0.1964 0.8544 0.9244
No log 5.7333 172 0.9792 0.0456 0.9792 0.9896
No log 5.8 174 1.0350 0.0267 1.0350 1.0173
No log 5.8667 176 0.9085 -0.0170 0.9085 0.9532
No log 5.9333 178 0.8836 -0.0984 0.8836 0.9400
No log 6.0 180 0.8740 -0.1964 0.8740 0.9349
No log 6.0667 182 0.9825 0.0424 0.9825 0.9912
No log 6.1333 184 1.2367 -0.1240 1.2367 1.1121
No log 6.2 186 1.1190 -0.0194 1.1190 1.0578
No log 6.2667 188 0.8765 -0.0204 0.8765 0.9362
No log 6.3333 190 0.8624 0.0282 0.8624 0.9287
No log 6.4 192 0.8962 0.1495 0.8962 0.9467
No log 6.4667 194 0.9475 0.1336 0.9475 0.9734
No log 6.5333 196 0.8796 -0.0583 0.8796 0.9379
No log 6.6 198 0.9032 -0.0195 0.9032 0.9504
No log 6.6667 200 1.0782 -0.0169 1.0782 1.0384
No log 6.7333 202 1.3070 -0.1504 1.3070 1.1432
No log 6.8 204 1.2396 -0.1201 1.2396 1.1134
No log 6.8667 206 0.9536 -0.0629 0.9536 0.9765
No log 6.9333 208 0.9035 -0.2030 0.9035 0.9505
No log 7.0 210 0.9254 -0.0661 0.9254 0.9620
No log 7.0667 212 0.8957 -0.1964 0.8957 0.9464
No log 7.1333 214 1.1513 -0.0899 1.1513 1.0730
No log 7.2 216 1.7123 -0.0941 1.7123 1.3086
No log 7.2667 218 1.6664 -0.0961 1.6664 1.2909
No log 7.3333 220 1.1821 -0.1281 1.1821 1.0873
No log 7.4 222 0.9350 -0.1126 0.9350 0.9670
No log 7.4667 224 0.9078 -0.2614 0.9078 0.9528
No log 7.5333 226 1.0054 -0.0799 1.0054 1.0027
No log 7.6 228 1.0427 -0.0818 1.0427 1.0211
No log 7.6667 230 0.9867 -0.0767 0.9867 0.9933
No log 7.7333 232 0.9434 -0.1176 0.9434 0.9713
No log 7.8 234 1.0317 -0.0408 1.0317 1.0157
No log 7.8667 236 1.0230 0.0871 1.0230 1.0114
No log 7.9333 238 0.9517 -0.1334 0.9517 0.9755
No log 8.0 240 0.9342 -0.1778 0.9342 0.9666
No log 8.0667 242 0.9912 0.0920 0.9912 0.9956
No log 8.1333 244 1.1872 -0.1186 1.1872 1.0896
No log 8.2 246 1.1951 -0.0482 1.1951 1.0932
No log 8.2667 248 1.0527 -0.0767 1.0527 1.0260
No log 8.3333 250 0.9337 -0.1682 0.9337 0.9663
No log 8.4 252 0.8813 -0.1230 0.8813 0.9388
No log 8.4667 254 0.9341 -0.1233 0.9341 0.9665
No log 8.5333 256 0.9722 -0.0753 0.9722 0.9860
No log 8.6 258 0.9778 -0.0287 0.9778 0.9888
No log 8.6667 260 0.9216 -0.1233 0.9216 0.9600
No log 8.7333 262 0.9465 -0.1172 0.9465 0.9729
No log 8.8 264 1.0151 -0.0766 1.0151 1.0075
No log 8.8667 266 0.9872 -0.1521 0.9872 0.9936
No log 8.9333 268 0.9835 -0.2123 0.9835 0.9917
No log 9.0 270 0.9629 -0.1967 0.9629 0.9813
No log 9.0667 272 0.9029 -0.2463 0.9029 0.9502
No log 9.1333 274 0.9363 -0.1184 0.9363 0.9676
No log 9.2 276 1.0463 -0.0471 1.0463 1.0229
No log 9.2667 278 1.0410 -0.0486 1.0410 1.0203
No log 9.3333 280 0.9331 -0.1197 0.9331 0.9660
No log 9.4 282 0.8938 -0.1675 0.8938 0.9454
No log 9.4667 284 0.8651 -0.1538 0.8651 0.9301
No log 9.5333 286 0.8710 -0.1094 0.8710 0.9333
No log 9.6 288 0.8624 -0.1094 0.8624 0.9286
No log 9.6667 290 0.9039 -0.1180 0.9039 0.9508
No log 9.7333 292 1.0829 -0.0571 1.0829 1.0406
No log 9.8 294 1.0773 0.0207 1.0773 1.0379
No log 9.8667 296 0.9372 0.0424 0.9372 0.9681
No log 9.9333 298 0.8418 -0.1163 0.8418 0.9175
No log 10.0 300 0.8371 -0.1535 0.8371 0.9149
No log 10.0667 302 0.8930 -0.0690 0.8930 0.9450
No log 10.1333 304 1.0831 0.0207 1.0831 1.0407
No log 10.2 306 1.2115 -0.1172 1.2115 1.1007
No log 10.2667 308 1.0826 0.1064 1.0826 1.0405
No log 10.3333 310 0.9013 -0.1594 0.9013 0.9494
No log 10.4 312 0.9007 -0.2383 0.9007 0.9491
No log 10.4667 314 0.8922 -0.2383 0.8922 0.9446
No log 10.5333 316 0.8593 -0.0660 0.8593 0.9270
No log 10.6 318 0.9841 0.1243 0.9841 0.9920
No log 10.6667 320 1.0927 -0.0930 1.0927 1.0453
No log 10.7333 322 0.9886 0.0676 0.9886 0.9943
No log 10.8 324 0.8610 -0.0644 0.8610 0.9279
No log 10.8667 326 0.8899 -0.2837 0.8899 0.9433
No log 10.9333 328 0.9194 -0.2966 0.9194 0.9589
No log 11.0 330 0.8834 -0.2935 0.8834 0.9399
No log 11.0667 332 0.8817 -0.1172 0.8817 0.9390
No log 11.1333 334 0.9682 -0.0442 0.9682 0.9840
No log 11.2 336 1.0179 -0.0532 1.0179 1.0089
No log 11.2667 338 0.9473 0.0786 0.9473 0.9733
No log 11.3333 340 0.8499 0.1047 0.8499 0.9219
No log 11.4 342 0.8357 -0.1158 0.8357 0.9142
No log 11.4667 344 0.8435 -0.1158 0.8435 0.9184
No log 11.5333 346 0.8478 -0.1158 0.8478 0.9208
No log 11.6 348 0.8550 -0.1094 0.8550 0.9247
No log 11.6667 350 0.8737 -0.0145 0.8737 0.9347
No log 11.7333 352 0.9262 0.0525 0.9262 0.9624
No log 11.8 354 0.9746 0.0786 0.9746 0.9872
No log 11.8667 356 0.9011 0.0867 0.9011 0.9493
No log 11.9333 358 0.7942 -0.0679 0.7942 0.8912
No log 12.0 360 0.7681 -0.0679 0.7681 0.8764
No log 12.0667 362 0.7747 -0.0160 0.7747 0.8802
No log 12.1333 364 0.8130 0.0768 0.8130 0.9017
No log 12.2 366 0.8191 -0.0152 0.8191 0.9050
No log 12.2667 368 0.8055 -0.0595 0.8055 0.8975
No log 12.3333 370 0.8480 0.1047 0.8480 0.9209
No log 12.4 372 0.9831 -0.0076 0.9831 0.9915
No log 12.4667 374 1.0526 -0.0532 1.0526 1.0260
No log 12.5333 376 0.9770 0.0362 0.9770 0.9885
No log 12.6 378 0.8673 0.0236 0.8673 0.9313
No log 12.6667 380 0.8776 -0.0999 0.8776 0.9368
No log 12.7333 382 0.9142 0.0091 0.9142 0.9561
No log 12.8 384 0.9066 0.0091 0.9066 0.9522
No log 12.8667 386 0.9275 0.0424 0.9275 0.9631
No log 12.9333 388 0.9466 0.0333 0.9466 0.9729
No log 13.0 390 0.9374 0.0333 0.9374 0.9682
No log 13.0667 392 0.9062 -0.1200 0.9062 0.9519
No log 13.1333 394 0.8810 -0.0132 0.8810 0.9386
No log 13.2 396 0.9168 -0.0699 0.9168 0.9575
No log 13.2667 398 1.0082 -0.0471 1.0082 1.0041
No log 13.3333 400 0.9832 0.0333 0.9832 0.9916
No log 13.4 402 0.8853 -0.1006 0.8853 0.9409
No log 13.4667 404 0.8772 -0.2173 0.8772 0.9366
No log 13.5333 406 0.9243 -0.2017 0.9243 0.9614
No log 13.6 408 0.9296 -0.1394 0.9296 0.9642
No log 13.6667 410 1.0185 0.0207 1.0185 1.0092
No log 13.7333 412 1.0486 -0.0955 1.0486 1.0240
No log 13.8 414 0.9734 -0.0474 0.9734 0.9866
No log 13.8667 416 0.9038 -0.1580 0.9038 0.9507
No log 13.9333 418 0.9007 -0.0567 0.9007 0.9491
No log 14.0 420 0.9071 -0.0599 0.9071 0.9524
No log 14.0667 422 0.9707 -0.0723 0.9707 0.9852
No log 14.1333 424 1.0298 0.0277 1.0298 1.0148
No log 14.2 426 1.0610 0.0277 1.0610 1.0301
No log 14.2667 428 1.1002 -0.0526 1.1002 1.0489
No log 14.3333 430 1.0287 0.0277 1.0287 1.0142
No log 14.4 432 0.8835 0.0525 0.8835 0.9400
No log 14.4667 434 0.8484 0.0953 0.8484 0.9211
No log 14.5333 436 0.8302 0.0159 0.8302 0.9111
No log 14.6 438 0.8014 -0.0725 0.8014 0.8952
No log 14.6667 440 0.8069 -0.1172 0.8069 0.8983
No log 14.7333 442 0.8372 -0.0660 0.8372 0.9150
No log 14.8 444 0.9264 0.0247 0.9264 0.9625
No log 14.8667 446 1.1038 -0.0146 1.1038 1.0506
No log 14.9333 448 1.1358 -0.0181 1.1358 1.0657
No log 15.0 450 1.0287 -0.0097 1.0287 1.0142
No log 15.0667 452 0.9875 -0.0441 0.9875 0.9937
No log 15.1333 454 0.9887 -0.0076 0.9887 0.9943
No log 15.2 456 0.9730 -0.0373 0.9730 0.9864
No log 15.2667 458 0.9612 -0.0355 0.9612 0.9804
No log 15.3333 460 0.9845 -0.0391 0.9845 0.9922
No log 15.4 462 1.1044 -0.0118 1.1044 1.0509
No log 15.4667 464 1.1943 -0.0583 1.1943 1.0928
No log 15.5333 466 1.1573 -0.0595 1.1573 1.0758
No log 15.6 468 0.9916 0.0260 0.9916 0.9958
No log 15.6667 470 0.8708 0.0424 0.8708 0.9331
No log 15.7333 472 0.8602 0.0095 0.8602 0.9275
No log 15.8 474 0.8722 -0.0316 0.8722 0.9339
No log 15.8667 476 0.8961 -0.0755 0.8961 0.9466
No log 15.9333 478 0.9250 -0.0336 0.9250 0.9617
No log 16.0 480 0.9745 -0.0031 0.9745 0.9872
No log 16.0667 482 1.0752 -0.0118 1.0752 1.0369
No log 16.1333 484 1.1454 -0.0899 1.1454 1.0702
No log 16.2 486 1.0185 -0.0157 1.0185 1.0092
No log 16.2667 488 0.9110 0.0333 0.9110 0.9544
No log 16.3333 490 0.8833 0.0041 0.8833 0.9398
No log 16.4 492 0.9172 0.0287 0.9172 0.9577
No log 16.4667 494 1.0068 -0.0182 1.0068 1.0034
No log 16.5333 496 1.0345 -0.0182 1.0345 1.0171
No log 16.6 498 0.9646 0.0316 0.9646 0.9821
0.292 16.6667 500 0.9358 0.0041 0.9358 0.9673
0.292 16.7333 502 0.9119 -0.0355 0.9119 0.9549
0.292 16.8 504 0.8703 -0.0731 0.8703 0.9329
0.292 16.8667 506 0.8794 -0.0316 0.8794 0.9378
0.292 16.9333 508 0.9127 -0.0355 0.9127 0.9553
0.292 17.0 510 0.9908 -0.0076 0.9908 0.9954

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k12_task3_organization

Finetuned
(4019)
this model