ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k14_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8483
  • Qwk: -0.0056
  • Mse: 0.8483
  • Rmse: 0.9210

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0571 2 3.8625 0.0035 3.8625 1.9653
No log 0.1143 4 2.3216 0.0494 2.3216 1.5237
No log 0.1714 6 3.3035 -0.0036 3.3035 1.8175
No log 0.2286 8 1.6256 -0.0453 1.6256 1.2750
No log 0.2857 10 1.0213 -0.0628 1.0213 1.0106
No log 0.3429 12 1.2204 -0.0736 1.2204 1.1047
No log 0.4 14 1.8874 -0.0512 1.8874 1.3738
No log 0.4571 16 1.6028 0.0 1.6028 1.2660
No log 0.5143 18 1.3958 0.0 1.3958 1.1814
No log 0.5714 20 1.0956 -0.0500 1.0956 1.0467
No log 0.6286 22 0.8349 -0.0309 0.8349 0.9137
No log 0.6857 24 0.8578 -0.0101 0.8578 0.9262
No log 0.7429 26 0.9328 -0.0345 0.9328 0.9658
No log 0.8 28 1.2204 -0.1019 1.2204 1.1047
No log 0.8571 30 1.1388 -0.0736 1.1388 1.0671
No log 0.9143 32 0.9146 0.0046 0.9146 0.9563
No log 0.9714 34 1.0353 -0.0385 1.0353 1.0175
No log 1.0286 36 1.2863 -0.0736 1.2863 1.1341
No log 1.0857 38 1.2132 -0.0712 1.2132 1.1014
No log 1.1429 40 0.8026 -0.0240 0.8026 0.8959
No log 1.2 42 0.7659 -0.0551 0.7659 0.8752
No log 1.2571 44 0.7585 -0.1067 0.7585 0.8709
No log 1.3143 46 0.8635 -0.1203 0.8635 0.9293
No log 1.3714 48 1.2352 -0.0207 1.2352 1.1114
No log 1.4286 50 1.2614 -0.0234 1.2614 1.1231
No log 1.4857 52 1.2242 -0.0234 1.2242 1.1064
No log 1.5429 54 1.1573 0.0048 1.1573 1.0758
No log 1.6 56 1.0258 0.0282 1.0258 1.0128
No log 1.6571 58 0.8413 -0.0643 0.8413 0.9172
No log 1.7143 60 0.9343 -0.0393 0.9343 0.9666
No log 1.7714 62 0.8595 0.0146 0.8595 0.9271
No log 1.8286 64 0.8763 0.0146 0.8763 0.9361
No log 1.8857 66 0.8164 -0.0170 0.8164 0.9036
No log 1.9429 68 0.8886 0.0588 0.8886 0.9427
No log 2.0 70 0.9968 0.0260 0.9968 0.9984
No log 2.0571 72 1.1771 -0.0936 1.1771 1.0849
No log 2.1143 74 0.7904 0.0355 0.7904 0.8891
No log 2.1714 76 0.8545 -0.1411 0.8545 0.9244
No log 2.2286 78 0.8852 -0.0557 0.8852 0.9408
No log 2.2857 80 0.9314 -0.0860 0.9314 0.9651
No log 2.3429 82 1.1707 -0.0409 1.1707 1.0820
No log 2.4 84 0.9497 -0.1173 0.9497 0.9745
No log 2.4571 86 1.0197 -0.0057 1.0197 1.0098
No log 2.5143 88 0.9408 -0.1202 0.9408 0.9699
No log 2.5714 90 0.9063 -0.0145 0.9063 0.9520
No log 2.6286 92 0.8878 -0.0170 0.8878 0.9422
No log 2.6857 94 0.8783 0.0690 0.8783 0.9372
No log 2.7429 96 0.9595 -0.0385 0.9595 0.9796
No log 2.8 98 0.9743 -0.0360 0.9743 0.9871
No log 2.8571 100 0.8716 -0.0082 0.8716 0.9336
No log 2.9143 102 1.3688 -0.1283 1.3688 1.1700
No log 2.9714 104 1.2721 -0.1944 1.2721 1.1279
No log 3.0286 106 0.8497 -0.0879 0.8497 0.9218
No log 3.0857 108 0.9643 -0.0551 0.9643 0.9820
No log 3.1429 110 0.9070 -0.1298 0.9070 0.9524
No log 3.2 112 0.7850 -0.0532 0.7850 0.8860
No log 3.2571 114 0.8156 0.0863 0.8156 0.9031
No log 3.3143 116 0.9347 0.0512 0.9347 0.9668
No log 3.3714 118 0.8022 -0.0532 0.8022 0.8956
No log 3.4286 120 0.8215 -0.1399 0.8215 0.9064
No log 3.4857 122 0.7706 -0.0513 0.7706 0.8779
No log 3.5429 124 0.8397 0.0999 0.8397 0.9164
No log 3.6 126 0.8330 0.0999 0.8330 0.9127
No log 3.6571 128 0.7777 -0.0826 0.7777 0.8819
No log 3.7143 130 0.9629 0.0092 0.9629 0.9813
No log 3.7714 132 0.9319 -0.0843 0.9319 0.9654
No log 3.8286 134 0.8281 0.0410 0.8281 0.9100
No log 3.8857 136 0.8691 0.1006 0.8691 0.9323
No log 3.9429 138 0.8145 0.0116 0.8145 0.9025
No log 4.0 140 0.8409 -0.0826 0.8409 0.9170
No log 4.0571 142 0.8332 -0.0082 0.8332 0.9128
No log 4.1143 144 0.9438 -0.0033 0.9438 0.9715
No log 4.1714 146 0.8505 0.0196 0.8505 0.9222
No log 4.2286 148 0.8539 -0.0826 0.8539 0.9241
No log 4.2857 150 0.8184 -0.0939 0.8184 0.9047
No log 4.3429 152 0.8425 0.0183 0.8425 0.9179
No log 4.4 154 0.8245 0.0863 0.8245 0.9080
No log 4.4571 156 0.8385 -0.0939 0.8385 0.9157
No log 4.5143 158 0.8646 -0.0949 0.8646 0.9298
No log 4.5714 160 0.9289 0.0183 0.9289 0.9638
No log 4.6286 162 0.8691 -0.0488 0.8691 0.9323
No log 4.6857 164 0.9115 -0.1647 0.9115 0.9547
No log 4.7429 166 0.8469 -0.0939 0.8469 0.9203
No log 4.8 168 0.8845 0.0562 0.8845 0.9405
No log 4.8571 170 0.8091 -0.0660 0.8091 0.8995
No log 4.9143 172 0.7741 0.0033 0.7741 0.8798
No log 4.9714 174 0.7578 0.0524 0.7578 0.8705
No log 5.0286 176 0.7700 -0.0096 0.7700 0.8775
No log 5.0857 178 0.8221 -0.0849 0.8221 0.9067
No log 5.1429 180 0.8578 0.0394 0.8578 0.9262
No log 5.2 182 0.8848 -0.0389 0.8848 0.9406
No log 5.2571 184 0.9463 -0.1062 0.9463 0.9728
No log 5.3143 186 0.9604 -0.1202 0.9604 0.9800
No log 5.3714 188 0.9322 -0.1155 0.9322 0.9655
No log 5.4286 190 0.8625 -0.1547 0.8625 0.9287
No log 5.4857 192 0.8622 -0.1106 0.8622 0.9285
No log 5.5429 194 0.9172 -0.0842 0.9172 0.9577
No log 5.6 196 0.8039 -0.0188 0.8039 0.8966
No log 5.6571 198 0.7633 -0.0499 0.7633 0.8737
No log 5.7143 200 0.7898 -0.0499 0.7898 0.8887
No log 5.7714 202 0.8356 -0.1473 0.8356 0.9141
No log 5.8286 204 0.8902 -0.1468 0.8902 0.9435
No log 5.8857 206 0.9376 -0.0446 0.9376 0.9683
No log 5.9429 208 0.9805 -0.2357 0.9805 0.9902
No log 6.0 210 0.9350 -0.2535 0.9350 0.9670
No log 6.0571 212 0.8708 -0.0976 0.8708 0.9332
No log 6.1143 214 0.9177 0.0017 0.9177 0.9580
No log 6.1714 216 0.8298 0.1311 0.8298 0.9110
No log 6.2286 218 0.8174 -0.1473 0.8174 0.9041
No log 6.2857 220 0.8683 -0.0881 0.8683 0.9318
No log 6.3429 222 0.8141 -0.1473 0.8141 0.9023
No log 6.4 224 0.8679 0.0341 0.8679 0.9316
No log 6.4571 226 0.8985 0.0277 0.8985 0.9479
No log 6.5143 228 0.9079 0.0277 0.9079 0.9528
No log 6.5714 230 0.8269 -0.1473 0.8269 0.9093
No log 6.6286 232 0.8213 -0.0520 0.8213 0.9063
No log 6.6857 234 0.9058 -0.0303 0.9058 0.9517
No log 6.7429 236 0.9194 0.0175 0.9194 0.9589
No log 6.8 238 0.9443 0.0196 0.9443 0.9718
No log 6.8571 240 0.9456 -0.1106 0.9456 0.9724
No log 6.9143 242 0.9749 -0.2227 0.9749 0.9874
No log 6.9714 244 0.9410 -0.1866 0.9410 0.9701
No log 7.0286 246 1.0672 -0.0787 1.0672 1.0330
No log 7.0857 248 1.1265 -0.2000 1.1265 1.0614
No log 7.1429 250 0.9368 0.0167 0.9368 0.9679
No log 7.2 252 0.8630 -0.1737 0.8630 0.9290
No log 7.2571 254 0.8411 -0.0449 0.8411 0.9171
No log 7.3143 256 0.8288 -0.1473 0.8288 0.9104
No log 7.3714 258 0.8161 -0.0065 0.8161 0.9034
No log 7.4286 260 0.8186 0.0807 0.8186 0.9048
No log 7.4857 262 0.8347 -0.1745 0.8347 0.9136
No log 7.5429 264 0.8653 -0.1151 0.8653 0.9302
No log 7.6 266 0.8749 -0.0051 0.8749 0.9354
No log 7.6571 268 0.8976 -0.1334 0.8976 0.9474
No log 7.7143 270 0.8745 -0.0573 0.8745 0.9352
No log 7.7714 272 0.8867 0.0650 0.8867 0.9417
No log 7.8286 274 0.8085 0.0394 0.8085 0.8992
No log 7.8857 276 0.7762 0.0513 0.7762 0.8810
No log 7.9429 278 0.8289 0.1097 0.8289 0.9104
No log 8.0 280 0.8677 0.0071 0.8677 0.9315
No log 8.0571 282 0.8143 0.1828 0.8143 0.9024
No log 8.1143 284 0.8336 -0.0465 0.8336 0.9130
No log 8.1714 286 0.8533 -0.1354 0.8533 0.9238
No log 8.2286 288 0.8327 0.0654 0.8327 0.9125
No log 8.2857 290 0.8882 -0.0209 0.8882 0.9425
No log 8.3429 292 1.0146 -0.0376 1.0146 1.0073
No log 8.4 294 0.8963 -0.0230 0.8963 0.9467
No log 8.4571 296 0.7113 0.1137 0.7113 0.8434
No log 8.5143 298 0.7938 0.1192 0.7938 0.8910
No log 8.5714 300 0.8087 0.0662 0.8087 0.8993
No log 8.6286 302 0.7567 0.0571 0.7567 0.8699
No log 8.6857 304 0.9490 -0.0425 0.9490 0.9742
No log 8.7429 306 1.2141 -0.0966 1.2141 1.1018
No log 8.8 308 1.1329 -0.0955 1.1329 1.0644
No log 8.8571 310 0.8772 0.0512 0.8772 0.9366
No log 8.9143 312 0.7134 0.1023 0.7134 0.8446
No log 8.9714 314 0.7924 -0.0322 0.7924 0.8902
No log 9.0286 316 0.8560 -0.1033 0.8560 0.9252
No log 9.0857 318 0.7937 0.0030 0.7937 0.8909
No log 9.1429 320 0.8839 0.0999 0.8839 0.9402
No log 9.2 322 0.8753 0.0152 0.8753 0.9356
No log 9.2571 324 0.8333 0.0934 0.8333 0.9129
No log 9.3143 326 0.8724 -0.2274 0.8724 0.9340
No log 9.3714 328 0.8163 -0.0892 0.8163 0.9035
No log 9.4286 330 0.7711 0.1304 0.7711 0.8781
No log 9.4857 332 0.7647 0.1249 0.7647 0.8745
No log 9.5429 334 0.7199 0.0454 0.7199 0.8485
No log 9.6 336 0.7497 -0.1001 0.7497 0.8659
No log 9.6571 338 0.8149 -0.0406 0.8149 0.9027
No log 9.7143 340 0.7831 -0.0406 0.7831 0.8849
No log 9.7714 342 0.7819 0.0 0.7819 0.8843
No log 9.8286 344 0.8510 0.0757 0.8510 0.9225
No log 9.8857 346 0.8498 -0.0326 0.8498 0.9218
No log 9.9429 348 0.8848 -0.1232 0.8848 0.9406
No log 10.0 350 0.8559 -0.0976 0.8559 0.9252
No log 10.0571 352 0.7809 0.0874 0.7809 0.8837
No log 10.1143 354 0.7638 0.0874 0.7638 0.8740
No log 10.1714 356 0.7685 0.0874 0.7685 0.8766
No log 10.2286 358 0.7937 0.1143 0.7937 0.8909
No log 10.2857 360 0.8468 0.1004 0.8468 0.9202
No log 10.3429 362 0.8175 0.0741 0.8175 0.9042
No log 10.4 364 0.7862 -0.0892 0.7862 0.8867
No log 10.4571 366 0.7791 -0.0892 0.7791 0.8826
No log 10.5143 368 0.7949 -0.0958 0.7949 0.8916
No log 10.5714 370 0.8541 0.0660 0.8541 0.9242
No log 10.6286 372 0.9010 0.0551 0.9010 0.9492
No log 10.6857 374 0.8412 0.0741 0.8412 0.9172
No log 10.7429 376 0.8211 -0.1330 0.8211 0.9061
No log 10.8 378 0.8278 -0.1266 0.8278 0.9099
No log 10.8571 380 0.8466 -0.1905 0.8466 0.9201
No log 10.9143 382 0.8354 0.0934 0.8354 0.9140
No log 10.9714 384 0.8536 0.1049 0.8536 0.9239
No log 11.0286 386 0.8382 0.1095 0.8382 0.9155
No log 11.0857 388 0.8134 0.0449 0.8134 0.9019
No log 11.1429 390 0.8505 -0.1979 0.8505 0.9222
No log 11.2 392 0.8478 -0.1151 0.8478 0.9207
No log 11.2571 394 0.8396 0.0622 0.8396 0.9163
No log 11.3143 396 0.8098 0.0732 0.8098 0.8999
No log 11.3714 398 0.7677 -0.0473 0.7677 0.8762
No log 11.4286 400 0.7463 -0.1001 0.7463 0.8639
No log 11.4857 402 0.7578 0.0031 0.7578 0.8705
No log 11.5429 404 0.7877 0.0874 0.7877 0.8875
No log 11.6 406 0.7770 0.0432 0.7770 0.8815
No log 11.6571 408 0.7774 0.0031 0.7774 0.8817
No log 11.7143 410 0.8185 -0.1905 0.8185 0.9047
No log 11.7714 412 0.8968 0.0359 0.8968 0.9470
No log 11.8286 414 0.9987 -0.0809 0.9987 0.9994
No log 11.8857 416 1.1188 -0.0936 1.1188 1.0577
No log 11.9429 418 1.0297 -0.0122 1.0297 1.0147
No log 12.0 420 0.8286 0.0741 0.8286 0.9103
No log 12.0571 422 0.7527 0.0662 0.7527 0.8676
No log 12.1143 424 0.7894 0.1192 0.7894 0.8885
No log 12.1714 426 0.7865 0.0094 0.7865 0.8869
No log 12.2286 428 0.9047 -0.0669 0.9047 0.9512
No log 12.2857 430 0.9553 -0.0408 0.9553 0.9774
No log 12.3429 432 0.9217 -0.0408 0.9217 0.9600
No log 12.4 434 0.8641 -0.0076 0.8641 0.9296
No log 12.4571 436 0.8503 -0.1472 0.8503 0.9221
No log 12.5143 438 0.8438 -0.0705 0.8438 0.9186
No log 12.5714 440 0.8641 0.0757 0.8641 0.9296
No log 12.6286 442 0.9297 0.0287 0.9297 0.9642
No log 12.6857 444 0.9403 0.0287 0.9403 0.9697
No log 12.7429 446 0.8810 0.1049 0.8810 0.9386
No log 12.8 448 0.8396 -0.0738 0.8396 0.9163
No log 12.8571 450 0.8873 0.0072 0.8873 0.9419
No log 12.9143 452 0.8385 -0.1538 0.8385 0.9157
No log 12.9714 454 0.8098 -0.0443 0.8098 0.8999
No log 13.0286 456 0.8514 -0.0095 0.8514 0.9227
No log 13.0857 458 0.9154 -0.0472 0.9154 0.9567
No log 13.1429 460 0.8909 -0.0047 0.8909 0.9439
No log 13.2 462 0.7952 -0.0912 0.7952 0.8918
No log 13.2571 464 0.7769 -0.1001 0.7769 0.8814
No log 13.3143 466 0.7477 -0.1001 0.7477 0.8647
No log 13.3714 468 0.7217 0.0460 0.7217 0.8495
No log 13.4286 470 0.7115 0.0 0.7115 0.8435
No log 13.4857 472 0.7144 0.0 0.7144 0.8452
No log 13.5429 474 0.7488 -0.1001 0.7488 0.8653
No log 13.6 476 0.7837 -0.1001 0.7837 0.8853
No log 13.6571 478 0.8196 -0.0473 0.8196 0.9053
No log 13.7143 480 0.8436 -0.0473 0.8436 0.9185
No log 13.7714 482 0.8417 -0.1678 0.8417 0.9174
No log 13.8286 484 0.8469 -0.0181 0.8469 0.9203
No log 13.8857 486 0.7866 -0.0881 0.7866 0.8869
No log 13.9429 488 0.7673 0.0496 0.7673 0.8760
No log 14.0 490 0.8044 0.1828 0.8044 0.8969
No log 14.0571 492 0.8034 0.1259 0.8034 0.8963
No log 14.1143 494 0.8047 0.1318 0.8047 0.8970
No log 14.1714 496 0.8090 0.0 0.8090 0.8995
No log 14.2286 498 0.8628 -0.0163 0.8628 0.9289
0.2799 14.2857 500 0.8897 0.0226 0.8897 0.9433
0.2799 14.3429 502 0.8421 0.0471 0.8421 0.9176
0.2799 14.4 504 0.8175 0.0173 0.8175 0.9042
0.2799 14.4571 506 0.7703 0.0155 0.7703 0.8777
0.2799 14.5143 508 0.7357 0.0496 0.7357 0.8578
0.2799 14.5714 510 0.8053 0.0538 0.8053 0.8974
0.2799 14.6286 512 0.8931 0.0424 0.8931 0.9451
0.2799 14.6857 514 0.8426 0.0525 0.8426 0.9179
0.2799 14.7429 516 0.7645 0.0031 0.7645 0.8744
0.2799 14.8 518 0.8087 -0.0284 0.8087 0.8993
0.2799 14.8571 520 0.8776 -0.1145 0.8776 0.9368
0.2799 14.9143 522 0.8681 -0.1530 0.8681 0.9317
0.2799 14.9714 524 0.8483 -0.0056 0.8483 0.9210

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k14_task3_organization

Finetuned
(4019)
this model