ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k8_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9412
  • Qwk: -0.1582
  • Mse: 0.9412
  • Rmse: 0.9702

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1 2 3.5026 -0.0354 3.5026 1.8715
No log 0.2 4 1.8202 0.0737 1.8202 1.3491
No log 0.3 6 1.1473 -0.1015 1.1473 1.0711
No log 0.4 8 1.0014 -0.0595 1.0014 1.0007
No log 0.5 10 0.7662 0.0159 0.7662 0.8753
No log 0.6 12 0.9721 -0.0606 0.9721 0.9859
No log 0.7 14 1.3685 0.0 1.3685 1.1698
No log 0.8 16 1.4891 0.0 1.4891 1.2203
No log 0.9 18 1.4558 0.0 1.4558 1.2066
No log 1.0 20 1.3516 -0.0247 1.3516 1.1626
No log 1.1 22 1.2995 -0.0221 1.2995 1.1399
No log 1.2 24 1.0748 -0.0961 1.0748 1.0367
No log 1.3 26 0.9893 -0.2364 0.9893 0.9946
No log 1.4 28 1.0093 -0.1609 1.0093 1.0047
No log 1.5 30 1.0983 -0.1292 1.0983 1.0480
No log 1.6 32 1.1615 -0.0457 1.1615 1.0777
No log 1.7 34 1.0375 -0.1287 1.0375 1.0186
No log 1.8 36 1.1011 -0.0712 1.1011 1.0494
No log 1.9 38 1.2093 -0.0457 1.2093 1.0997
No log 2.0 40 1.1977 -0.0149 1.1977 1.0944
No log 2.1 42 1.0196 -0.0285 1.0196 1.0098
No log 2.2 44 0.9584 -0.0861 0.9584 0.9790
No log 2.3 46 0.9547 0.0676 0.9547 0.9771
No log 2.4 48 1.0532 -0.1281 1.0532 1.0263
No log 2.5 50 1.1994 -0.0695 1.1994 1.0952
No log 2.6 52 1.2841 0.0152 1.2841 1.1332
No log 2.7 54 1.5708 0.0690 1.5708 1.2533
No log 2.8 56 1.1618 -0.0628 1.1618 1.0779
No log 2.9 58 1.1594 -0.0628 1.1594 1.0768
No log 3.0 60 0.9128 0.0409 0.9128 0.9554
No log 3.1 62 0.8510 0.0909 0.8510 0.9225
No log 3.2 64 0.8099 0.1291 0.8099 0.8999
No log 3.3 66 0.8015 0.0549 0.8015 0.8953
No log 3.4 68 1.0633 0.0735 1.0633 1.0311
No log 3.5 70 0.9835 0.1145 0.9835 0.9917
No log 3.6 72 0.8169 0.0257 0.8169 0.9038
No log 3.7 74 1.0930 0.0707 1.0930 1.0455
No log 3.8 76 1.2563 -0.0353 1.2563 1.1209
No log 3.9 78 1.1650 -0.0288 1.1650 1.0794
No log 4.0 80 0.8573 -0.0560 0.8573 0.9259
No log 4.1 82 0.8698 -0.0560 0.8698 0.9326
No log 4.2 84 1.0710 -0.0862 1.0710 1.0349
No log 4.3 86 1.2867 -0.0620 1.2867 1.1343
No log 4.4 88 1.0423 -0.1224 1.0423 1.0209
No log 4.5 90 1.0527 -0.1214 1.0527 1.0260
No log 4.6 92 1.2328 -0.0835 1.2328 1.1103
No log 4.7 94 1.1152 -0.1224 1.1152 1.0560
No log 4.8 96 0.9693 -0.1255 0.9693 0.9845
No log 4.9 98 0.8492 -0.0560 0.8492 0.9215
No log 5.0 100 0.8381 -0.0595 0.8381 0.9155
No log 5.1 102 1.2223 0.0252 1.2223 1.1056
No log 5.2 104 1.6883 0.0502 1.6883 1.2993
No log 5.3 106 1.2619 -0.0098 1.2619 1.1233
No log 5.4 108 0.8875 -0.0718 0.8875 0.9421
No log 5.5 110 0.9048 -0.1187 0.9048 0.9512
No log 5.6 112 1.0763 -0.0862 1.0763 1.0374
No log 5.7 114 0.9536 -0.2046 0.9536 0.9765
No log 5.8 116 0.8951 -0.0578 0.8951 0.9461
No log 5.9 118 0.9248 -0.0541 0.9248 0.9617
No log 6.0 120 0.9798 -0.2534 0.9798 0.9899
No log 6.1 122 0.9897 -0.2511 0.9897 0.9948
No log 6.2 124 0.9480 -0.0612 0.9480 0.9736
No log 6.3 126 1.0230 -0.2035 1.0230 1.0114
No log 6.4 128 1.2056 -0.1837 1.2056 1.0980
No log 6.5 130 1.1397 -0.1490 1.1397 1.0676
No log 6.6 132 0.9422 -0.1100 0.9422 0.9707
No log 6.7 134 0.9429 -0.1893 0.9429 0.9711
No log 6.8 136 0.8953 -0.0560 0.8953 0.9462
No log 6.9 138 1.1732 -0.1154 1.1732 1.0832
No log 7.0 140 1.6023 -0.0422 1.6023 1.2658
No log 7.1 142 1.4553 -0.0916 1.4553 1.2064
No log 7.2 144 0.9760 -0.2035 0.9760 0.9879
No log 7.3 146 0.9071 -0.0578 0.9071 0.9524
No log 7.4 148 0.8928 -0.1116 0.8928 0.9449
No log 7.5 150 1.0996 -0.0930 1.0996 1.0486
No log 7.6 152 1.2240 0.0217 1.2240 1.1063
No log 7.7 154 1.0041 -0.1645 1.0041 1.0020
No log 7.8 156 0.9370 -0.2072 0.9370 0.9680
No log 7.9 158 0.9184 -0.2077 0.9184 0.9583
No log 8.0 160 1.0713 -0.1988 1.0713 1.0350
No log 8.1 162 1.1124 -0.1597 1.1124 1.0547
No log 8.2 164 0.9672 -0.2425 0.9672 0.9835
No log 8.3 166 0.8725 -0.1088 0.8725 0.9341
No log 8.4 168 0.8361 -0.0595 0.8361 0.9144
No log 8.5 170 1.0275 -0.1586 1.0275 1.0137
No log 8.6 172 1.1823 -0.0012 1.1823 1.0873
No log 8.7 174 0.8862 -0.1212 0.8862 0.9414
No log 8.8 176 0.7712 0.0436 0.7712 0.8782
No log 8.9 178 0.7894 -0.1191 0.7894 0.8885
No log 9.0 180 0.9922 -0.0331 0.9922 0.9961
No log 9.1 182 1.0712 -0.0677 1.0712 1.0350
No log 9.2 184 1.0506 -0.0617 1.0506 1.0250
No log 9.3 186 1.0644 -0.0885 1.0644 1.0317
No log 9.4 188 0.9209 -0.2131 0.9209 0.9596
No log 9.5 190 0.9623 -0.1468 0.9623 0.9810
No log 9.6 192 0.9380 -0.1951 0.9380 0.9685
No log 9.7 194 1.1543 -0.0563 1.1543 1.0744
No log 9.8 196 1.3673 -0.0348 1.3673 1.1693
No log 9.9 198 1.0504 -0.0916 1.0504 1.0249
No log 10.0 200 0.9123 -0.1633 0.9123 0.9552
No log 10.1 202 0.8989 -0.1176 0.8989 0.9481
No log 10.2 204 0.9981 -0.2429 0.9981 0.9990
No log 10.3 206 1.2304 -0.0306 1.2304 1.1093
No log 10.4 208 1.5039 -0.1219 1.5039 1.2263
No log 10.5 210 1.1995 -0.0925 1.1995 1.0952
No log 10.6 212 0.8954 -0.1158 0.8954 0.9463
No log 10.7 214 0.9667 -0.2435 0.9667 0.9832
No log 10.8 216 0.9782 -0.2733 0.9782 0.9890
No log 10.9 218 0.9365 -0.2022 0.9365 0.9677
No log 11.0 220 1.2423 -0.1279 1.2423 1.1146
No log 11.1 222 1.6440 -0.0928 1.6440 1.2822
No log 11.2 224 1.5584 -0.0443 1.5584 1.2484
No log 11.3 226 1.1543 -0.1620 1.1543 1.0744
No log 11.4 228 0.9952 -0.3389 0.9952 0.9976
No log 11.5 230 0.9416 -0.1163 0.9416 0.9704
No log 11.6 232 0.9120 -0.1153 0.9120 0.9550
No log 11.7 234 0.9046 -0.1158 0.9046 0.9511
No log 11.8 236 0.9275 -0.2652 0.9275 0.9631
No log 11.9 238 0.9861 -0.1261 0.9861 0.9930
No log 12.0 240 1.0100 -0.0504 1.0100 1.0050
No log 12.1 242 0.8956 -0.1715 0.8956 0.9464
No log 12.2 244 0.8937 -0.2607 0.8937 0.9453
No log 12.3 246 1.0031 -0.2410 1.0031 1.0015
No log 12.4 248 1.0798 -0.1230 1.0798 1.0392
No log 12.5 250 1.0779 -0.0892 1.0779 1.0382
No log 12.6 252 0.9967 -0.1967 0.9967 0.9984
No log 12.7 254 0.9724 -0.2374 0.9724 0.9861
No log 12.8 256 0.9973 -0.1579 0.9973 0.9987
No log 12.9 258 0.9656 -0.2021 0.9656 0.9826
No log 13.0 260 0.9118 -0.2086 0.9118 0.9549
No log 13.1 262 0.9495 -0.2046 0.9495 0.9744
No log 13.2 264 1.0760 -0.0575 1.0760 1.0373
No log 13.3 266 1.0499 -0.0575 1.0499 1.0246
No log 13.4 268 0.9664 -0.1212 0.9664 0.9831
No log 13.5 270 0.8843 -0.1633 0.8843 0.9404
No log 13.6 272 0.9172 -0.1609 0.9172 0.9577
No log 13.7 274 0.9631 -0.1214 0.9631 0.9814
No log 13.8 276 0.9830 -0.1586 0.9830 0.9914
No log 13.9 278 0.9952 -0.0606 0.9952 0.9976
No log 14.0 280 0.9557 -0.1221 0.9557 0.9776
No log 14.1 282 0.9409 -0.1219 0.9409 0.9700
No log 14.2 284 0.9049 -0.2046 0.9049 0.9512
No log 14.3 286 0.9440 -0.1614 0.9440 0.9716
No log 14.4 288 0.9998 -0.0845 0.9998 0.9999
No log 14.5 290 0.8791 -0.0743 0.8791 0.9376
No log 14.6 292 0.8398 -0.0560 0.8398 0.9164
No log 14.7 294 0.8449 -0.0560 0.8449 0.9192
No log 14.8 296 0.8973 -0.1187 0.8973 0.9473
No log 14.9 298 1.0849 -0.0285 1.0849 1.0416
No log 15.0 300 1.1211 -0.0031 1.1211 1.0588
No log 15.1 302 1.0140 -0.1939 1.0140 1.0070
No log 15.2 304 0.9459 -0.2046 0.9459 0.9726
No log 15.3 306 0.9099 -0.1187 0.9099 0.9539
No log 15.4 308 0.9415 -0.1589 0.9415 0.9703
No log 15.5 310 1.0367 -0.0331 1.0367 1.0182
No log 15.6 312 1.0741 -0.0658 1.0741 1.0364
No log 15.7 314 1.0254 -0.1576 1.0254 1.0126
No log 15.8 316 0.9891 -0.1589 0.9891 0.9945
No log 15.9 318 0.9698 -0.1217 0.9698 0.9848
No log 16.0 320 0.9176 -0.1618 0.9176 0.9579
No log 16.1 322 0.8602 -0.1168 0.8602 0.9275
No log 16.2 324 0.8961 -0.0541 0.8961 0.9466
No log 16.3 326 0.8608 -0.0560 0.8608 0.9278
No log 16.4 328 0.8665 -0.1650 0.8665 0.9308
No log 16.5 330 1.1711 0.0204 1.1711 1.0822
No log 16.6 332 1.3266 -0.0655 1.3266 1.1518
No log 16.7 334 1.0675 -0.0013 1.0675 1.0332
No log 16.8 336 0.8066 -0.1730 0.8066 0.8981
No log 16.9 338 0.8963 -0.1547 0.8963 0.9467
No log 17.0 340 0.9596 -0.1905 0.9596 0.9796
No log 17.1 342 0.9077 -0.1074 0.9077 0.9527
No log 17.2 344 0.8609 -0.1172 0.8609 0.9279
No log 17.3 346 1.0690 -0.0885 1.0690 1.0339
No log 17.4 348 1.2151 -0.0348 1.2151 1.1023
No log 17.5 350 1.1578 -0.0359 1.1578 1.0760
No log 17.6 352 0.9981 -0.0877 0.9981 0.9990
No log 17.7 354 0.8945 -0.0799 0.8945 0.9458
No log 17.8 356 0.8647 -0.2582 0.8647 0.9299
No log 17.9 358 0.8821 -0.2116 0.8821 0.9392
No log 18.0 360 0.8902 -0.2557 0.8902 0.9435
No log 18.1 362 0.9280 -0.1593 0.9280 0.9633
No log 18.2 364 0.9464 -0.0892 0.9464 0.9728
No log 18.3 366 0.9376 -0.0595 0.9376 0.9683
No log 18.4 368 0.9688 -0.0269 0.9688 0.9843
No log 18.5 370 0.9139 -0.0143 0.9139 0.9560
No log 18.6 372 0.8543 -0.2511 0.8543 0.9243
No log 18.7 374 0.8189 -0.0660 0.8189 0.9049
No log 18.8 376 0.8922 -0.2981 0.8922 0.9445
No log 18.9 378 1.0758 -0.0551 1.0758 1.0372
No log 19.0 380 1.1655 -0.0597 1.1655 1.0796
No log 19.1 382 1.0843 -0.0905 1.0843 1.0413
No log 19.2 384 1.0446 -0.0563 1.0446 1.0220
No log 19.3 386 0.9440 -0.0845 0.9440 0.9716
No log 19.4 388 0.8654 -0.0799 0.8654 0.9303
No log 19.5 390 0.9496 -0.0194 0.9496 0.9745
No log 19.6 392 1.1093 -0.0677 1.1093 1.0532
No log 19.7 394 1.2243 -0.0716 1.2243 1.1065
No log 19.8 396 1.2044 -0.0716 1.2044 1.0974
No log 19.9 398 1.0601 -0.0638 1.0601 1.0296
No log 20.0 400 0.9895 -0.0905 0.9895 0.9947
No log 20.1 402 0.9565 -0.0854 0.9565 0.9780
No log 20.2 404 0.9880 -0.0513 0.9880 0.9940
No log 20.3 406 1.0616 -0.0586 1.0616 1.0304
No log 20.4 408 1.1592 -0.0937 1.1592 1.0767
No log 20.5 410 1.1100 -0.0608 1.1100 1.0536
No log 20.6 412 1.0408 -0.0597 1.0408 1.0202
No log 20.7 414 0.9801 -0.1579 0.9801 0.9900
No log 20.8 416 0.9882 -0.1232 0.9882 0.9941
No log 20.9 418 0.9625 -0.1582 0.9625 0.9811
No log 21.0 420 0.9330 -0.1597 0.9330 0.9659
No log 21.1 422 0.8876 -0.2468 0.8876 0.9421
No log 21.2 424 0.8717 -0.2468 0.8717 0.9337
No log 21.3 426 0.9103 -0.1221 0.9103 0.9541
No log 21.4 428 0.9525 0.0067 0.9525 0.9760
No log 21.5 430 0.9543 -0.0253 0.9543 0.9769
No log 21.6 432 0.8592 -0.0852 0.8592 0.9269
No log 21.7 434 0.8422 -0.1168 0.8422 0.9177
No log 21.8 436 0.9348 -0.0902 0.9348 0.9668
No log 21.9 438 0.9489 -0.1396 0.9489 0.9741
No log 22.0 440 0.9700 -0.2059 0.9700 0.9849
No log 22.1 442 1.1880 -0.0648 1.1880 1.0900
No log 22.2 444 1.2025 -0.0648 1.2025 1.0966
No log 22.3 446 1.1379 -0.0586 1.1379 1.0667
No log 22.4 448 1.0812 -0.0899 1.0812 1.0398
No log 22.5 450 1.0265 -0.1582 1.0265 1.0131
No log 22.6 452 0.9811 -0.2033 0.9811 0.9905
No log 22.7 454 0.9735 -0.2429 0.9735 0.9867
No log 22.8 456 0.9783 -0.1217 0.9783 0.9891
No log 22.9 458 0.9374 -0.1214 0.9374 0.9682
No log 23.0 460 0.9052 -0.1618 0.9052 0.9514
No log 23.1 462 0.8893 -0.2557 0.8893 0.9430
No log 23.2 464 0.9056 -0.1628 0.9056 0.9516
No log 23.3 466 0.9426 -0.1217 0.9426 0.9709
No log 23.4 468 1.0597 -0.0563 1.0597 1.0294
No log 23.5 470 1.0874 -0.0586 1.0874 1.0428
No log 23.6 472 1.1167 -0.0638 1.1167 1.0567
No log 23.7 474 1.0897 -0.0638 1.0897 1.0439
No log 23.8 476 1.0070 -0.0899 1.0070 1.0035
No log 23.9 478 0.9267 -0.2046 0.9267 0.9627
No log 24.0 480 0.9255 -0.1633 0.9255 0.9620
No log 24.1 482 0.9527 -0.1224 0.9527 0.9761
No log 24.2 484 0.9809 -0.0575 0.9809 0.9904
No log 24.3 486 1.0134 -0.0301 1.0134 1.0067
No log 24.4 488 0.9738 -0.0575 0.9738 0.9868
No log 24.5 490 0.9255 -0.0837 0.9255 0.9620
No log 24.6 492 0.9213 -0.1614 0.9213 0.9598
No log 24.7 494 0.9347 -0.1614 0.9347 0.9668
No log 24.8 496 0.9750 -0.0471 0.9750 0.9874
No log 24.9 498 1.0362 -0.0175 1.0362 1.0179
0.2367 25.0 500 1.1786 -0.0597 1.1786 1.0856
0.2367 25.1 502 1.2949 -0.0677 1.2949 1.1379
0.2367 25.2 504 1.3204 -0.0398 1.3204 1.1491
0.2367 25.3 506 1.1872 -0.0359 1.1872 1.0896
0.2367 25.4 508 1.0374 -0.0586 1.0374 1.0185
0.2367 25.5 510 0.9482 -0.1618 0.9482 0.9737
0.2367 25.6 512 0.9158 -0.1638 0.9158 0.9570
0.2367 25.7 514 0.8874 -0.1184 0.8874 0.9420
0.2367 25.8 516 0.8889 -0.2891 0.8889 0.9428
0.2367 25.9 518 0.9412 -0.1582 0.9412 0.9702

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k8_task3_organization

Finetuned
(4019)
this model