ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k6_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7396
  • Qwk: 0.0
  • Mse: 0.7396
  • Rmse: 0.8600

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1333 2 3.6362 -0.0047 3.6362 1.9069
No log 0.2667 4 2.1430 0.0247 2.1430 1.4639
No log 0.4 6 1.7879 0.0398 1.7879 1.3371
No log 0.5333 8 1.1515 -0.0234 1.1515 1.0731
No log 0.6667 10 1.1172 0.0610 1.1172 1.0570
No log 0.8 12 0.9793 0.1153 0.9793 0.9896
No log 0.9333 14 1.0696 -0.0207 1.0696 1.0342
No log 1.0667 16 0.9217 -0.0617 0.9217 0.9601
No log 1.2 18 0.8039 -0.0331 0.8039 0.8966
No log 1.3333 20 0.8656 -0.0504 0.8656 0.9304
No log 1.4667 22 0.8276 -0.0842 0.8276 0.9097
No log 1.6 24 0.8536 -0.1265 0.8536 0.9239
No log 1.7333 26 1.0474 0.0338 1.0474 1.0234
No log 1.8667 28 1.2419 0.0 1.2419 1.1144
No log 2.0 30 1.3791 0.0 1.3791 1.1744
No log 2.1333 32 1.2663 -0.0247 1.2663 1.1253
No log 2.2667 34 1.2035 0.0083 1.2035 1.0971
No log 2.4 36 0.8726 -0.0878 0.8726 0.9341
No log 2.5333 38 0.7571 -0.1220 0.7571 0.8701
No log 2.6667 40 0.7289 -0.1216 0.7289 0.8538
No log 2.8 42 0.8517 -0.1249 0.8517 0.9229
No log 2.9333 44 1.1466 -0.0982 1.1466 1.0708
No log 3.0667 46 1.3335 0.0083 1.3335 1.1548
No log 3.2 48 1.4895 -0.0247 1.4895 1.2205
No log 3.3333 50 1.4622 -0.0247 1.4622 1.2092
No log 3.4667 52 1.3291 0.0 1.3291 1.1529
No log 3.6 54 1.2690 0.0 1.2690 1.1265
No log 3.7333 56 1.1171 -0.0234 1.1171 1.0569
No log 3.8667 58 0.9108 -0.0200 0.9108 0.9543
No log 4.0 60 0.8024 -0.0790 0.8024 0.8958
No log 4.1333 62 0.7801 -0.0739 0.7801 0.8832
No log 4.2667 64 0.7806 -0.0739 0.7806 0.8835
No log 4.4 66 0.8013 -0.0739 0.8013 0.8952
No log 4.5333 68 0.7833 -0.0739 0.7833 0.8851
No log 4.6667 70 0.8425 -0.1244 0.8425 0.9179
No log 4.8 72 1.0115 0.0111 1.0115 1.0057
No log 4.9333 74 1.1528 0.0493 1.1528 1.0737
No log 5.0667 76 1.2981 -0.0193 1.2981 1.1394
No log 5.2 78 1.0448 -0.0049 1.0448 1.0222
No log 5.3333 80 0.7543 -0.1241 0.7543 0.8685
No log 5.4667 82 0.7242 0.0555 0.7242 0.8510
No log 5.6 84 0.7500 -0.0551 0.7500 0.8660
No log 5.7333 86 0.7645 -0.1168 0.7645 0.8743
No log 5.8667 88 0.9029 0.0642 0.9029 0.9502
No log 6.0 90 0.8932 -0.0008 0.8932 0.9451
No log 6.1333 92 0.8333 0.0191 0.8333 0.9129
No log 6.2667 94 0.8175 -0.0264 0.8175 0.9041
No log 6.4 96 0.8194 -0.0096 0.8194 0.9052
No log 6.5333 98 0.8447 0.0095 0.8447 0.9191
No log 6.6667 100 0.9161 0.0068 0.9161 0.9571
No log 6.8 102 0.9162 0.0129 0.9162 0.9572
No log 6.9333 104 0.9473 -0.0672 0.9473 0.9733
No log 7.0667 106 0.9932 -0.1194 0.9932 0.9966
No log 7.2 108 0.9956 -0.0059 0.9956 0.9978
No log 7.3333 110 0.9080 0.0040 0.9080 0.9529
No log 7.4667 112 0.8915 -0.0778 0.8915 0.9442
No log 7.6 114 0.8783 -0.1106 0.8783 0.9372
No log 7.7333 116 0.8470 -0.0385 0.8470 0.9203
No log 7.8667 118 0.7679 -0.0662 0.7679 0.8763
No log 8.0 120 0.8481 -0.0371 0.8481 0.9209
No log 8.1333 122 0.9753 0.0545 0.9753 0.9876
No log 8.2667 124 0.8629 -0.0425 0.8629 0.9290
No log 8.4 126 0.7736 0.0 0.7736 0.8795
No log 8.5333 128 0.9080 -0.0408 0.9080 0.9529
No log 8.6667 130 0.9466 -0.1054 0.9466 0.9729
No log 8.8 132 0.8200 0.0260 0.8200 0.9055
No log 8.9333 134 0.7494 0.0556 0.7494 0.8657
No log 9.0667 136 0.7500 0.0556 0.7500 0.8661
No log 9.2 138 0.7970 -0.0329 0.7970 0.8928
No log 9.3333 140 0.8578 0.1034 0.8578 0.9262
No log 9.4667 142 0.8495 0.0268 0.8495 0.9217
No log 9.6 144 0.7929 0.0179 0.7929 0.8904
No log 9.7333 146 0.8041 0.0151 0.8041 0.8967
No log 9.8667 148 0.8317 -0.0116 0.8317 0.9120
No log 10.0 150 0.8436 0.0315 0.8436 0.9185
No log 10.1333 152 0.8500 0.1123 0.8500 0.9220
No log 10.2667 154 0.8125 -0.0204 0.8125 0.9014
No log 10.4 156 0.7720 -0.0428 0.7720 0.8786
No log 10.5333 158 0.7545 0.0602 0.7545 0.8686
No log 10.6667 160 0.7478 0.0524 0.7478 0.8648
No log 10.8 162 0.7419 0.0555 0.7419 0.8613
No log 10.9333 164 0.7051 0.0 0.7051 0.8397
No log 11.0667 166 0.7025 -0.0033 0.7025 0.8382
No log 11.2 168 0.7478 0.0318 0.7478 0.8648
No log 11.3333 170 0.7074 0.0436 0.7074 0.8411
No log 11.4667 172 0.7640 -0.0406 0.7640 0.8741
No log 11.6 174 0.7730 0.0064 0.7730 0.8792
No log 11.7333 176 0.7184 0.0 0.7184 0.8476
No log 11.8667 178 0.7123 0.0479 0.7123 0.8440
No log 12.0 180 0.9114 0.0362 0.9114 0.9547
No log 12.1333 182 0.9412 -0.0076 0.9412 0.9702
No log 12.2667 184 0.7451 0.0183 0.7451 0.8632
No log 12.4 186 0.7025 0.0 0.7025 0.8381
No log 12.5333 188 0.7434 -0.0033 0.7434 0.8622
No log 12.6667 190 0.7637 -0.1074 0.7637 0.8739
No log 12.8 192 0.7249 0.0587 0.7249 0.8514
No log 12.9333 194 0.7147 0.1552 0.7147 0.8454
No log 13.0667 196 0.7330 0.2024 0.7330 0.8562
No log 13.2 198 0.8155 0.0240 0.8155 0.9031
No log 13.3333 200 0.9518 -0.0583 0.9518 0.9756
No log 13.4667 202 0.9827 -0.0253 0.9827 0.9913
No log 13.6 204 0.9431 -0.0334 0.9431 0.9711
No log 13.7333 206 0.8472 0.0260 0.8472 0.9205
No log 13.8667 208 0.8137 -0.0958 0.8137 0.9021
No log 14.0 210 0.7698 -0.0520 0.7698 0.8774
No log 14.1333 212 0.7388 -0.0520 0.7388 0.8595
No log 14.2667 214 0.7286 0.0 0.7286 0.8536
No log 14.4 216 0.7092 0.0 0.7092 0.8421
No log 14.5333 218 0.7063 0.0 0.7063 0.8404
No log 14.6667 220 0.7158 0.0 0.7158 0.8460
No log 14.8 222 0.7335 0.0 0.7335 0.8564
No log 14.9333 224 0.7419 0.0033 0.7419 0.8613
No log 15.0667 226 0.7672 0.0033 0.7672 0.8759
No log 15.2 228 0.7865 0.0640 0.7865 0.8869
No log 15.3333 230 0.8132 0.0155 0.8132 0.9018
No log 15.4667 232 0.8572 0.0733 0.8572 0.9259
No log 15.6 234 0.8215 0.0640 0.8215 0.9064
No log 15.7333 236 0.7573 0.0615 0.7573 0.8702
No log 15.8667 238 0.7416 0.0973 0.7416 0.8612
No log 16.0 240 0.7624 -0.0271 0.7624 0.8732
No log 16.1333 242 0.8321 0.1126 0.8321 0.9122
No log 16.2667 244 0.8373 0.0718 0.8373 0.9150
No log 16.4 246 0.8078 0.1991 0.8078 0.8988
No log 16.5333 248 0.7918 0.0652 0.7918 0.8898
No log 16.6667 250 0.7959 0.0640 0.7959 0.8921
No log 16.8 252 0.7896 0.0615 0.7896 0.8886
No log 16.9333 254 0.7753 0.0602 0.7753 0.8805
No log 17.0667 256 0.7516 0.0033 0.7516 0.8670
No log 17.2 258 0.7228 0.0602 0.7228 0.8502
No log 17.3333 260 0.7305 0.0602 0.7305 0.8547
No log 17.4667 262 0.7767 0.1135 0.7767 0.8813
No log 17.6 264 0.7946 0.1137 0.7946 0.8914
No log 17.7333 266 0.7705 0.1137 0.7705 0.8778
No log 17.8667 268 0.7509 0.0602 0.7509 0.8665
No log 18.0 270 0.7485 0.0033 0.7485 0.8652
No log 18.1333 272 0.7831 0.1137 0.7831 0.8849
No log 18.2667 274 0.8479 0.0673 0.8479 0.9208
No log 18.4 276 0.8883 0.0734 0.8883 0.9425
No log 18.5333 278 0.8466 0.0664 0.8466 0.9201
No log 18.6667 280 0.8017 0.1562 0.8017 0.8954
No log 18.8 282 0.7680 0.0556 0.7680 0.8764
No log 18.9333 284 0.7664 0.0513 0.7664 0.8754
No log 19.0667 286 0.7820 0.1080 0.7820 0.8843
No log 19.2 288 0.8016 0.1611 0.8016 0.8953
No log 19.3333 290 0.7674 0.0602 0.7674 0.8760
No log 19.4667 292 0.7531 0.0602 0.7531 0.8678
No log 19.6 294 0.7193 0.0602 0.7193 0.8481
No log 19.7333 296 0.7052 0.0 0.7052 0.8398
No log 19.8667 298 0.7183 0.1137 0.7183 0.8475
No log 20.0 300 0.7667 0.0652 0.7667 0.8756
No log 20.1333 302 0.8189 0.0279 0.8189 0.9049
No log 20.2667 304 0.8379 0.1121 0.8379 0.9154
No log 20.4 306 0.7858 0.0652 0.7858 0.8865
No log 20.5333 308 0.7291 0.0587 0.7291 0.8539
No log 20.6667 310 0.7465 0.0914 0.7465 0.8640
No log 20.8 312 0.7583 0.0615 0.7583 0.8708
No log 20.9333 314 0.7457 0.0587 0.7457 0.8635
No log 21.0667 316 0.6977 0.0555 0.6977 0.8353
No log 21.2 318 0.6961 0.0555 0.6961 0.8343
No log 21.3333 320 0.7212 0.1143 0.7212 0.8492
No log 21.4667 322 0.7657 0.1024 0.7657 0.8750
No log 21.6 324 0.8059 0.0116 0.8059 0.8977
No log 21.7333 326 0.8801 -0.1059 0.8801 0.9381
No log 21.8667 328 0.9111 0.0107 0.9111 0.9545
No log 22.0 330 0.8767 -0.0425 0.8767 0.9363
No log 22.1333 332 0.8092 0.0260 0.8092 0.8996
No log 22.2667 334 0.7619 0.0064 0.7619 0.8729
No log 22.4 336 0.7312 0.0587 0.7312 0.8551
No log 22.5333 338 0.7279 0.0587 0.7279 0.8532
No log 22.6667 340 0.7263 0.0587 0.7263 0.8522
No log 22.8 342 0.7571 0.0640 0.7571 0.8701
No log 22.9333 344 0.7618 0.0640 0.7618 0.8728
No log 23.0667 346 0.7460 0.2071 0.7460 0.8637
No log 23.2 348 0.7405 0.2368 0.7405 0.8605
No log 23.3333 350 0.7387 0.1928 0.7387 0.8595
No log 23.4667 352 0.7374 0.2044 0.7374 0.8587
No log 23.6 354 0.7546 0.1641 0.7546 0.8687
No log 23.7333 356 0.7448 0.1192 0.7448 0.8630
No log 23.8667 358 0.6924 0.0 0.6924 0.8321
No log 24.0 360 0.6740 0.0 0.6740 0.8210
No log 24.1333 362 0.6711 0.1082 0.6711 0.8192
No log 24.2667 364 0.6711 0.0555 0.6711 0.8192
No log 24.4 366 0.6901 0.0 0.6901 0.8307
No log 24.5333 368 0.7820 0.0662 0.7820 0.8843
No log 24.6667 370 0.8597 0.0789 0.8597 0.9272
No log 24.8 372 0.8511 0.0418 0.8511 0.9226
No log 24.9333 374 0.7799 0.1514 0.7799 0.8831
No log 25.0667 376 0.7768 0.2275 0.7768 0.8814
No log 25.2 378 0.7919 0.1408 0.7919 0.8899
No log 25.3333 380 0.8549 0.1112 0.8549 0.9246
No log 25.4667 382 0.9384 0.1389 0.9384 0.9687
No log 25.6 384 0.9103 0.0240 0.9103 0.9541
No log 25.7333 386 0.7986 -0.0144 0.7986 0.8937
No log 25.8667 388 0.7010 0.1143 0.7010 0.8373
No log 26.0 390 0.6731 0.0555 0.6731 0.8204
No log 26.1333 392 0.6856 0.0914 0.6856 0.8280
No log 26.2667 394 0.6927 0.0967 0.6927 0.8323
No log 26.4 396 0.6932 0.0524 0.6932 0.8326
No log 26.5333 398 0.7031 0.0587 0.7031 0.8385
No log 26.6667 400 0.7284 0.1137 0.7284 0.8534
No log 26.8 402 0.7290 0.1139 0.7290 0.8538
No log 26.9333 404 0.7210 0.0587 0.7210 0.8491
No log 27.0667 406 0.7045 0.0587 0.7045 0.8393
No log 27.2 408 0.6741 0.0524 0.6741 0.8211
No log 27.3333 410 0.6882 0.0914 0.6882 0.8296
No log 27.4667 412 0.7096 0.1423 0.7096 0.8423
No log 27.6 414 0.7171 0.1362 0.7171 0.8468
No log 27.7333 416 0.7333 0.1576 0.7333 0.8563
No log 27.8667 418 0.7449 0.0122 0.7449 0.8630
No log 28.0 420 0.7153 0.0587 0.7153 0.8458
No log 28.1333 422 0.6912 0.0555 0.6912 0.8314
No log 28.2667 424 0.6978 0.1639 0.6978 0.8353
No log 28.4 426 0.7317 0.0587 0.7317 0.8554
No log 28.5333 428 0.7632 0.0640 0.7632 0.8736
No log 28.6667 430 0.7599 0.1135 0.7599 0.8717
No log 28.8 432 0.7460 0.1137 0.7460 0.8637
No log 28.9333 434 0.7402 0.0587 0.7402 0.8604
No log 29.0667 436 0.7357 0.0587 0.7357 0.8577
No log 29.2 438 0.7343 0.0 0.7343 0.8569
No log 29.3333 440 0.7499 0.0602 0.7499 0.8659
No log 29.4667 442 0.7878 0.0094 0.7878 0.8876
No log 29.6 444 0.7790 0.0587 0.7790 0.8826
No log 29.7333 446 0.7544 0.0587 0.7544 0.8686
No log 29.8667 448 0.7196 0.0555 0.7196 0.8483
No log 30.0 450 0.6901 0.0555 0.6901 0.8307
No log 30.1333 452 0.6932 0.0555 0.6932 0.8326
No log 30.2667 454 0.7286 0.0555 0.7286 0.8536
No log 30.4 456 0.7776 0.0064 0.7776 0.8818
No log 30.5333 458 0.8213 0.0640 0.8213 0.9062
No log 30.6667 460 0.8194 0.0640 0.8194 0.9052
No log 30.8 462 0.7678 0.0587 0.7678 0.8762
No log 30.9333 464 0.6993 0.0 0.6993 0.8363
No log 31.0667 466 0.6851 0.0479 0.6851 0.8277
No log 31.2 468 0.6940 0.0479 0.6940 0.8331
No log 31.3333 470 0.7105 0.0479 0.7105 0.8429
No log 31.4667 472 0.7242 0.0 0.7242 0.8510
No log 31.6 474 0.7241 0.0 0.7241 0.8510
No log 31.7333 476 0.7133 0.0 0.7133 0.8446
No log 31.8667 478 0.7043 0.0 0.7043 0.8392
No log 32.0 480 0.6932 0.0555 0.6932 0.8326
No log 32.1333 482 0.6967 0.0524 0.6967 0.8347
No log 32.2667 484 0.7168 0.0 0.7168 0.8466
No log 32.4 486 0.7390 0.0602 0.7390 0.8596
No log 32.5333 488 0.7718 0.0673 0.7718 0.8785
No log 32.6667 490 0.7752 0.0673 0.7752 0.8805
No log 32.8 492 0.7757 0.1626 0.7757 0.8807
No log 32.9333 494 0.7708 0.0602 0.7708 0.8780
No log 33.0667 496 0.7536 0.0587 0.7536 0.8681
No log 33.2 498 0.7319 0.0 0.7319 0.8555
0.3061 33.3333 500 0.6987 0.0 0.6987 0.8359
0.3061 33.4667 502 0.6822 0.0 0.6822 0.8260
0.3061 33.6 504 0.6776 0.0 0.6776 0.8232
0.3061 33.7333 506 0.6827 0.0 0.6827 0.8263
0.3061 33.8667 508 0.7123 0.0 0.7123 0.8440
0.3061 34.0 510 0.7296 0.0587 0.7296 0.8542
0.3061 34.1333 512 0.7359 0.0587 0.7359 0.8578
0.3061 34.2667 514 0.7177 0.0 0.7177 0.8472
0.3061 34.4 516 0.7034 0.0 0.7034 0.8387
0.3061 34.5333 518 0.7088 0.0 0.7088 0.8419
0.3061 34.6667 520 0.7516 0.0587 0.7516 0.8669
0.3061 34.8 522 0.7890 0.0587 0.7890 0.8883
0.3061 34.9333 524 0.8010 0.0587 0.8010 0.8950
0.3061 35.0667 526 0.7730 0.0068 0.7730 0.8792
0.3061 35.2 528 0.7210 0.0555 0.7210 0.8491
0.3061 35.3333 530 0.6897 0.0555 0.6897 0.8305
0.3061 35.4667 532 0.6895 0.0555 0.6895 0.8304
0.3061 35.6 534 0.6977 0.0555 0.6977 0.8353
0.3061 35.7333 536 0.7145 0.0479 0.7145 0.8453
0.3061 35.8667 538 0.7445 0.0479 0.7445 0.8628
0.3061 36.0 540 0.7706 0.0541 0.7706 0.8778
0.3061 36.1333 542 0.7854 0.0587 0.7854 0.8862
0.3061 36.2667 544 0.7741 0.0541 0.7741 0.8799
0.3061 36.4 546 0.7442 -0.0033 0.7442 0.8627
0.3061 36.5333 548 0.7165 -0.0033 0.7165 0.8465
0.3061 36.6667 550 0.6958 0.0506 0.6958 0.8342
0.3061 36.8 552 0.6928 0.0555 0.6928 0.8323
0.3061 36.9333 554 0.7092 0.0555 0.7092 0.8421
0.3061 37.0667 556 0.7408 0.0555 0.7408 0.8607
0.3061 37.2 558 0.7610 0.0602 0.7610 0.8723
0.3061 37.3333 560 0.7754 0.1135 0.7754 0.8806
0.3061 37.4667 562 0.7712 0.1135 0.7712 0.8782
0.3061 37.6 564 0.7922 0.1133 0.7922 0.8901
0.3061 37.7333 566 0.7813 0.1141 0.7813 0.8839
0.3061 37.8667 568 0.7532 0.1143 0.7532 0.8678
0.3061 38.0 570 0.7295 0.0555 0.7295 0.8541
0.3061 38.1333 572 0.7388 0.0555 0.7388 0.8595
0.3061 38.2667 574 0.7567 0.0033 0.7567 0.8699
0.3061 38.4 576 0.7869 0.1137 0.7869 0.8871
0.3061 38.5333 578 0.7765 0.0602 0.7765 0.8812
0.3061 38.6667 580 0.7574 0.0 0.7574 0.8703
0.3061 38.8 582 0.7335 0.0 0.7335 0.8564
0.3061 38.9333 584 0.7234 0.0 0.7234 0.8506
0.3061 39.0667 586 0.7299 0.0 0.7299 0.8544
0.3061 39.2 588 0.7396 0.0 0.7396 0.8600

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k6_task3_organization

Finetuned
(4019)
this model