ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k16_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8382
  • Qwk: -0.1905
  • Mse: 0.8382
  • Rmse: 0.9155

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 3.3980 0.0062 3.3980 1.8434
No log 0.1 4 1.9708 0.0325 1.9708 1.4038
No log 0.15 6 2.1479 0.0247 2.1479 1.4656
No log 0.2 8 1.5226 0.0014 1.5226 1.2340
No log 0.25 10 0.8722 0.0714 0.8722 0.9339
No log 0.3 12 0.8362 -0.0264 0.8362 0.9144
No log 0.35 14 1.2668 -0.0744 1.2668 1.1255
No log 0.4 16 2.4290 -0.0305 2.4290 1.5585
No log 0.45 18 2.2510 -0.0299 2.2510 1.5003
No log 0.5 20 1.2395 -0.0500 1.2395 1.1133
No log 0.55 22 1.0327 0.0735 1.0327 1.0162
No log 0.6 24 0.9211 0.1374 0.9211 0.9597
No log 0.65 26 0.9035 0.0651 0.9035 0.9505
No log 0.7 28 0.9468 0.0196 0.9468 0.9730
No log 0.75 30 1.0873 -0.0193 1.0873 1.0427
No log 0.8 32 1.0973 -0.0490 1.0973 1.0475
No log 0.85 34 1.1334 -0.0490 1.1334 1.0646
No log 0.9 36 1.1097 -0.0736 1.1097 1.0534
No log 0.95 38 1.1091 -0.0736 1.1091 1.0532
No log 1.0 40 0.9710 -0.1289 0.9710 0.9854
No log 1.05 42 0.8849 -0.0518 0.8849 0.9407
No log 1.1 44 1.0090 -0.0398 1.0090 1.0045
No log 1.15 46 1.0081 -0.0695 1.0081 1.0040
No log 1.2 48 1.0240 -0.0149 1.0240 1.0119
No log 1.25 50 1.1165 -0.0457 1.1165 1.0566
No log 1.3 52 1.7256 -0.0234 1.7256 1.3136
No log 1.35 54 2.0120 0.0213 2.0120 1.4184
No log 1.4 56 1.5820 -0.0234 1.5820 1.2578
No log 1.45 58 0.9980 0.0176 0.9980 0.9990
No log 1.5 60 0.7930 -0.1236 0.7930 0.8905
No log 1.55 62 0.7941 -0.1738 0.7941 0.8911
No log 1.6 64 0.8189 -0.1251 0.8189 0.9049
No log 1.65 66 0.8699 -0.0218 0.8699 0.9327
No log 1.7 68 1.0543 -0.0468 1.0543 1.0268
No log 1.75 70 1.3181 -0.0234 1.3181 1.1481
No log 1.8 72 1.3516 -0.0234 1.3516 1.1626
No log 1.85 74 1.2640 -0.0234 1.2640 1.1243
No log 1.9 76 1.1986 -0.0207 1.1986 1.0948
No log 1.95 78 1.4452 -0.0234 1.4452 1.2022
No log 2.0 80 1.5705 -0.0234 1.5705 1.2532
No log 2.05 82 1.7436 0.0224 1.7436 1.3205
No log 2.1 84 1.8182 0.1052 1.8182 1.3484
No log 2.15 86 1.4007 0.0522 1.4007 1.1835
No log 2.2 88 0.8701 -0.0200 0.8701 0.9328
No log 2.25 90 0.8599 0.1107 0.8599 0.9273
No log 2.3 92 1.1321 0.0552 1.1321 1.0640
No log 2.35 94 1.4306 -0.0169 1.4306 1.1961
No log 2.4 96 1.7726 -0.0035 1.7726 1.3314
No log 2.45 98 1.6684 -0.0197 1.6684 1.2916
No log 2.5 100 1.0109 0.1846 1.0109 1.0054
No log 2.55 102 0.8130 0.0956 0.8130 0.9017
No log 2.6 104 0.9424 0.1896 0.9424 0.9708
No log 2.65 106 1.8840 0.0144 1.8840 1.3726
No log 2.7 108 1.9333 0.0100 1.9333 1.3904
No log 2.75 110 1.4854 -0.0465 1.4854 1.2188
No log 2.8 112 0.9600 0.0498 0.9600 0.9798
No log 2.85 114 0.9603 0.0515 0.9603 0.9800
No log 2.9 116 1.1712 -0.0345 1.1712 1.0822
No log 2.95 118 1.0519 0.0046 1.0519 1.0256
No log 3.0 120 0.8397 0.0442 0.8397 0.9164
No log 3.05 122 0.8996 0.0711 0.8996 0.9485
No log 3.1 124 1.0893 0.0061 1.0893 1.0437
No log 3.15 126 0.9432 0.0576 0.9432 0.9712
No log 3.2 128 0.7313 0.0414 0.7313 0.8551
No log 3.25 130 0.7540 -0.0062 0.7540 0.8683
No log 3.3 132 1.2220 0.0059 1.2220 1.1054
No log 3.35 134 1.4430 -0.0028 1.4430 1.2012
No log 3.4 136 0.9168 0.1190 0.9168 0.9575
No log 3.45 138 0.8572 0.1004 0.8572 0.9259
No log 3.5 140 0.7929 0.0700 0.7929 0.8904
No log 3.55 142 0.9007 0.1147 0.9007 0.9490
No log 3.6 144 1.0464 0.0481 1.0464 1.0229
No log 3.65 146 0.7777 0.0732 0.7777 0.8819
No log 3.7 148 0.7598 0.1199 0.7598 0.8717
No log 3.75 150 0.8024 0.1047 0.8024 0.8958
No log 3.8 152 0.8399 0.1387 0.8399 0.9165
No log 3.85 154 0.8404 0.1239 0.8404 0.9167
No log 3.9 156 0.7832 0.1096 0.7832 0.8850
No log 3.95 158 0.7622 0.1249 0.7622 0.8730
No log 4.0 160 0.7702 0.0323 0.7702 0.8776
No log 4.05 162 0.8633 0.0421 0.8633 0.9291
No log 4.1 164 0.9760 0.0519 0.9760 0.9879
No log 4.15 166 0.8730 0.1277 0.8730 0.9343
No log 4.2 168 0.7690 0.0930 0.7690 0.8769
No log 4.25 170 0.7488 0.0543 0.7488 0.8653
No log 4.3 172 0.7522 0.0089 0.7522 0.8673
No log 4.35 174 0.7425 0.1815 0.7425 0.8617
No log 4.4 176 0.7914 0.0956 0.7914 0.8896
No log 4.45 178 0.7199 0.0030 0.7199 0.8485
No log 4.5 180 0.7335 -0.0449 0.7335 0.8565
No log 4.55 182 0.7820 0.0338 0.7820 0.8843
No log 4.6 184 0.8086 -0.0113 0.8086 0.8992
No log 4.65 186 0.8141 -0.0163 0.8141 0.9023
No log 4.7 188 0.7803 -0.0446 0.7803 0.8834
No log 4.75 190 0.7989 0.0791 0.7989 0.8938
No log 4.8 192 0.8082 0.0341 0.8082 0.8990
No log 4.85 194 0.8035 0.0056 0.8035 0.8964
No log 4.9 196 0.8846 0.0682 0.8846 0.9405
No log 4.95 198 0.8392 0.0056 0.8392 0.9161
No log 5.0 200 1.0779 -0.0228 1.0779 1.0382
No log 5.05 202 1.5791 -0.0932 1.5791 1.2566
No log 5.1 204 1.2910 0.0265 1.2910 1.1362
No log 5.15 206 0.8407 -0.0963 0.8407 0.9169
No log 5.2 208 0.8325 0.1525 0.8325 0.9124
No log 5.25 210 0.8928 0.1078 0.8928 0.9449
No log 5.3 212 0.7808 0.1865 0.7808 0.8837
No log 5.35 214 0.8344 0.0040 0.8344 0.9135
No log 5.4 216 0.7982 -0.0322 0.7982 0.8934
No log 5.45 218 0.7558 0.1080 0.7558 0.8694
No log 5.5 220 0.8101 -0.0259 0.8101 0.9000
No log 5.55 222 0.7780 0.0089 0.7780 0.8820
No log 5.6 224 0.7654 -0.0118 0.7654 0.8749
No log 5.65 226 0.8463 -0.0359 0.8463 0.9199
No log 5.7 228 0.7806 0.0926 0.7806 0.8835
No log 5.75 230 0.8093 0.0116 0.8093 0.8996
No log 5.8 232 0.8101 0.0598 0.8101 0.9000
No log 5.85 234 0.8136 0.1028 0.8136 0.9020
No log 5.9 236 0.8274 -0.1599 0.8274 0.9096
No log 5.95 238 0.8164 -0.1599 0.8164 0.9035
No log 6.0 240 0.7986 0.0031 0.7986 0.8936
No log 6.05 242 0.7832 0.0031 0.7832 0.8850
No log 6.1 244 0.8358 -0.0718 0.8358 0.9142
No log 6.15 246 0.9068 -0.0818 0.9068 0.9523
No log 6.2 248 0.8324 -0.1187 0.8324 0.9124
No log 6.25 250 0.7585 -0.0033 0.7585 0.8709
No log 6.3 252 0.7695 -0.0493 0.7695 0.8772
No log 6.35 254 0.7876 -0.0541 0.7876 0.8875
No log 6.4 256 0.9736 -0.0094 0.9736 0.9867
No log 6.45 258 1.0420 -0.0486 1.0420 1.0208
No log 6.5 260 0.8687 -0.0274 0.8687 0.9321
No log 6.55 262 0.8075 0.0089 0.8075 0.8986
No log 6.6 264 0.8112 0.0089 0.8112 0.9007
No log 6.65 266 0.8079 0.0116 0.8079 0.8988
No log 6.7 268 0.8365 -0.0113 0.8365 0.9146
No log 6.75 270 1.3180 -0.0353 1.3180 1.1480
No log 6.8 272 1.7056 -0.0187 1.7056 1.3060
No log 6.85 274 1.3606 -0.0366 1.3606 1.1664
No log 6.9 276 0.8024 0.0732 0.8024 0.8958
No log 6.95 278 0.8567 -0.0798 0.8567 0.9256
No log 7.0 280 0.8583 -0.0881 0.8583 0.9264
No log 7.05 282 0.7949 -0.0493 0.7949 0.8916
No log 7.1 284 0.7457 -0.0030 0.7457 0.8636
No log 7.15 286 0.8710 0.0793 0.8710 0.9333
No log 7.2 288 0.8866 0.0719 0.8866 0.9416
No log 7.25 290 0.7406 0.0436 0.7406 0.8606
No log 7.3 292 0.7577 -0.0493 0.7577 0.8705
No log 7.35 294 0.7999 -0.0428 0.7999 0.8944
No log 7.4 296 0.7702 0.0061 0.7702 0.8776
No log 7.45 298 0.7817 0.0394 0.7817 0.8842
No log 7.5 300 0.9046 -0.0359 0.9046 0.9511
No log 7.55 302 0.9869 -0.0157 0.9869 0.9934
No log 7.6 304 0.9858 0.0224 0.9858 0.9928
No log 7.65 306 0.8846 -0.0723 0.8846 0.9405
No log 7.7 308 0.7979 -0.0532 0.7979 0.8932
No log 7.75 310 0.7985 -0.0532 0.7985 0.8936
No log 7.8 312 0.8067 -0.1531 0.8067 0.8981
No log 7.85 314 0.8657 -0.0686 0.8657 0.9304
No log 7.9 316 1.0092 -0.0892 1.0092 1.0046
No log 7.95 318 1.0384 -0.0563 1.0384 1.0190
No log 8.0 320 0.9122 -0.0408 0.9122 0.9551
No log 8.05 322 0.8270 -0.1054 0.8270 0.9094
No log 8.1 324 0.8227 -0.0030 0.8227 0.9070
No log 8.15 326 0.8334 -0.1599 0.8334 0.9129
No log 8.2 328 0.9043 0.0016 0.9043 0.9510
No log 8.25 330 1.0001 -0.0877 1.0001 1.0000
No log 8.3 332 0.9901 -0.0885 0.9901 0.9950
No log 8.35 334 0.8743 0.0123 0.8743 0.9351
No log 8.4 336 0.8361 -0.1033 0.8361 0.9144
No log 8.45 338 0.8453 -0.1100 0.8453 0.9194
No log 8.5 340 0.9085 -0.0031 0.9085 0.9531
No log 8.55 342 0.9264 -0.0118 0.9264 0.9625
No log 8.6 344 0.8497 -0.0743 0.8497 0.9218
No log 8.65 346 0.7951 0.0031 0.7951 0.8917
No log 8.7 348 0.8061 -0.1329 0.8061 0.8978
No log 8.75 350 0.7857 -0.1468 0.7857 0.8864
No log 8.8 352 0.7941 -0.1100 0.7941 0.8911
No log 8.85 354 0.9570 -0.0500 0.9570 0.9783
No log 8.9 356 1.2193 -0.0385 1.2193 1.1042
No log 8.95 358 1.3514 -0.0411 1.3514 1.1625
No log 9.0 360 1.2131 -0.0657 1.2131 1.1014
No log 9.05 362 0.9464 -0.0486 0.9464 0.9728
No log 9.1 364 0.8173 0.0357 0.8173 0.9040
No log 9.15 366 0.7713 -0.0902 0.7713 0.8782
No log 9.2 368 0.7620 0.0031 0.7620 0.8729
No log 9.25 370 0.7680 -0.0541 0.7680 0.8764
No log 9.3 372 0.7864 -0.1153 0.7864 0.8868
No log 9.35 374 0.7980 -0.1153 0.7980 0.8933
No log 9.4 376 0.8394 -0.2148 0.8394 0.9162
No log 9.45 378 0.8774 -0.1194 0.8774 0.9367
No log 9.5 380 0.8790 -0.1200 0.8790 0.9375
No log 9.55 382 0.9113 -0.0097 0.9113 0.9546
No log 9.6 384 0.9038 0.0287 0.9038 0.9507
No log 9.65 386 0.8245 0.0525 0.8245 0.9080
No log 9.7 388 0.7643 0.0821 0.7643 0.8742
No log 9.75 390 0.7435 0.0 0.7435 0.8623
No log 9.8 392 0.7552 0.0585 0.7552 0.8690
No log 9.85 394 0.7445 0.0543 0.7445 0.8628
No log 9.9 396 0.8140 0.1495 0.8140 0.9022
No log 9.95 398 0.9243 -0.0545 0.9243 0.9614
No log 10.0 400 0.8510 -0.0391 0.8510 0.9225
No log 10.05 402 0.7739 0.1292 0.7739 0.8797
No log 10.1 404 0.7748 0.0978 0.7748 0.8802
No log 10.15 406 0.7883 0.0660 0.7883 0.8879
No log 10.2 408 0.9134 -0.0471 0.9134 0.9557
No log 10.25 410 0.9339 -0.0486 0.9339 0.9664
No log 10.3 412 0.8896 -0.0746 0.8896 0.9432
No log 10.35 414 0.8519 -0.0711 0.8519 0.9230
No log 10.4 416 0.8272 -0.1585 0.8272 0.9095
No log 10.45 418 0.7821 -0.0118 0.7821 0.8844
No log 10.5 420 0.7753 0.0338 0.7753 0.8805
No log 10.55 422 0.8923 -0.0778 0.8923 0.9446
No log 10.6 424 1.0182 -0.0918 1.0182 1.0091
No log 10.65 426 0.9766 -0.0211 0.9766 0.9882
No log 10.7 428 0.8406 0.0118 0.8406 0.9168
No log 10.75 430 0.7498 0.0375 0.7498 0.8659
No log 10.8 432 0.7419 0.0471 0.7419 0.8613
No log 10.85 434 0.7562 -0.0062 0.7562 0.8696
No log 10.9 436 0.8437 -0.1140 0.8437 0.9185
No log 10.95 438 0.8780 -0.0735 0.8780 0.9370
No log 11.0 440 0.9007 -0.0359 0.9007 0.9491
No log 11.05 442 0.9371 -0.1172 0.9371 0.9681
No log 11.1 444 0.9479 -0.0845 0.9479 0.9736
No log 11.15 446 0.9066 -0.0341 0.9066 0.9522
No log 11.2 448 0.7922 -0.0643 0.7922 0.8900
No log 11.25 450 0.7904 -0.0628 0.7904 0.8891
No log 11.3 452 0.9154 -0.0359 0.9154 0.9567
No log 11.35 454 1.0463 -0.0892 1.0463 1.0229
No log 11.4 456 0.9952 -0.1172 0.9952 0.9976
No log 11.45 458 0.9092 -0.0303 0.9092 0.9535
No log 11.5 460 0.8757 -0.0218 0.8757 0.9358
No log 11.55 462 0.8109 -0.0091 0.8109 0.9005
No log 11.6 464 0.8004 -0.0091 0.8004 0.8947
No log 11.65 466 0.7887 -0.0091 0.7887 0.8881
No log 11.7 468 0.8094 -0.1594 0.8094 0.8997
No log 11.75 470 0.8003 -0.0578 0.8003 0.8946
No log 11.8 472 0.7954 -0.0560 0.7954 0.8919
No log 11.85 474 0.8275 -0.0578 0.8275 0.9097
No log 11.9 476 0.9049 0.0876 0.9049 0.9512
No log 11.95 478 1.0113 -0.1232 1.0113 1.0056
No log 12.0 480 0.9903 -0.1232 0.9903 0.9952
No log 12.05 482 0.8355 0.1144 0.8355 0.9141
No log 12.1 484 0.7947 -0.0902 0.7947 0.8915
No log 12.15 486 0.8742 -0.1399 0.8742 0.9350
No log 12.2 488 0.8610 -0.1399 0.8610 0.9279
No log 12.25 490 0.7783 0.0030 0.7783 0.8822
No log 12.3 492 0.8300 0.1379 0.8300 0.9110
No log 12.35 494 1.0134 -0.1240 1.0134 1.0067
No log 12.4 496 0.9888 -0.1238 0.9888 0.9944
No log 12.45 498 0.8623 0.0424 0.8623 0.9286
0.3465 12.5 500 0.7570 -0.0091 0.7570 0.8700
0.3465 12.55 502 0.7478 -0.1010 0.7478 0.8648
0.3465 12.6 504 0.7602 -0.0385 0.7602 0.8719
0.3465 12.65 506 0.7454 -0.0493 0.7454 0.8634
0.3465 12.7 508 0.7589 -0.0118 0.7589 0.8711
0.3465 12.75 510 0.7957 0.1048 0.7957 0.8920
0.3465 12.8 512 0.7653 0.1048 0.7653 0.8748
0.3465 12.85 514 0.7351 -0.0091 0.7351 0.8574
0.3465 12.9 516 0.7279 -0.0062 0.7279 0.8532
0.3465 12.95 518 0.7210 -0.0595 0.7210 0.8491
0.3465 13.0 520 0.7590 0.1146 0.7590 0.8712
0.3465 13.05 522 0.9014 -0.0486 0.9014 0.9494
0.3465 13.1 524 0.9250 -0.0563 0.9250 0.9618
0.3465 13.15 526 0.9077 -0.0837 0.9077 0.9527
0.3465 13.2 528 0.8545 0.0068 0.8545 0.9244
0.3465 13.25 530 0.8323 0.0562 0.8323 0.9123
0.3465 13.3 532 0.8061 0.1001 0.8061 0.8978
0.3465 13.35 534 0.7672 0.0338 0.7672 0.8759
0.3465 13.4 536 0.7637 -0.1094 0.7637 0.8739
0.3465 13.45 538 0.7662 -0.1033 0.7662 0.8753
0.3465 13.5 540 0.7705 -0.0595 0.7705 0.8778
0.3465 13.55 542 0.7955 -0.0218 0.7955 0.8919
0.3465 13.6 544 0.8735 0.0831 0.8735 0.9346
0.3465 13.65 546 0.9734 -0.0551 0.9734 0.9866
0.3465 13.7 548 0.9584 -0.0539 0.9584 0.9790
0.3465 13.75 550 0.8761 0.0793 0.8761 0.9360
0.3465 13.8 552 0.7811 -0.0195 0.7811 0.8838
0.3465 13.85 554 0.8208 -0.0195 0.8208 0.9060
0.3465 13.9 556 0.9144 -0.0306 0.9144 0.9562
0.3465 13.95 558 0.8757 -0.0111 0.8757 0.9358
0.3465 14.0 560 0.8382 -0.1905 0.8382 0.9155

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k16_task3_organization

Finetuned
(4019)
this model