ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k13_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8513
  • Qwk: -0.0690
  • Mse: 0.8513
  • Rmse: 0.9227

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0606 2 3.6736 0.0048 3.6736 1.9167
No log 0.1212 4 1.7469 0.0943 1.7469 1.3217
No log 0.1818 6 1.0541 0.0912 1.0541 1.0267
No log 0.2424 8 1.1992 0.0349 1.1992 1.0951
No log 0.3030 10 1.0208 0.0623 1.0208 1.0104
No log 0.3636 12 0.7325 0.0714 0.7325 0.8559
No log 0.4242 14 0.7086 0.0807 0.7086 0.8418
No log 0.4848 16 0.9212 0.0873 0.9212 0.9598
No log 0.5455 18 0.8457 0.0346 0.8457 0.9196
No log 0.6061 20 0.9808 -0.0133 0.9808 0.9904
No log 0.6667 22 1.2189 -0.0234 1.2189 1.1040
No log 0.7273 24 1.3556 -0.0247 1.3556 1.1643
No log 0.7879 26 1.4957 -0.0247 1.4957 1.2230
No log 0.8485 28 1.3123 -0.0247 1.3123 1.1455
No log 0.9091 30 0.9156 -0.0617 0.9156 0.9569
No log 0.9697 32 0.7125 -0.0035 0.7125 0.8441
No log 1.0303 34 0.8743 0.0786 0.8743 0.9350
No log 1.0909 36 1.3676 -0.0751 1.3676 1.1694
No log 1.1515 38 1.2008 -0.0744 1.2008 1.0958
No log 1.2121 40 0.7763 0.1691 0.7763 0.8811
No log 1.2727 42 0.6841 -0.0035 0.6841 0.8271
No log 1.3333 44 0.8009 0.2180 0.8009 0.8949
No log 1.3939 46 1.2216 -0.0751 1.2216 1.1053
No log 1.4545 48 0.9902 -0.0301 0.9902 0.9951
No log 1.5152 50 0.7109 -0.0571 0.7109 0.8431
No log 1.5758 52 0.7452 0.0 0.7452 0.8633
No log 1.6364 54 0.7071 0.0 0.7071 0.8409
No log 1.6970 56 0.7002 0.0 0.7002 0.8368
No log 1.7576 58 1.1330 -0.0479 1.1330 1.0644
No log 1.8182 60 1.8708 0.0 1.8708 1.3678
No log 1.8788 62 2.3058 0.0591 2.3058 1.5185
No log 1.9394 64 2.2367 0.0 2.2367 1.4956
No log 2.0 66 2.1252 -0.0247 2.1252 1.4578
No log 2.0606 68 1.7669 -0.0234 1.7669 1.3292
No log 2.1212 70 1.2728 -0.0490 1.2728 1.1282
No log 2.1818 72 0.8634 0.0287 0.8634 0.9292
No log 2.2424 74 0.7659 0.1318 0.7659 0.8752
No log 2.3030 76 0.8384 0.1395 0.8384 0.9156
No log 2.3636 78 1.2176 -0.1011 1.2176 1.1034
No log 2.4242 80 2.1481 -0.0278 2.1481 1.4656
No log 2.4848 82 2.6658 0.0298 2.6658 1.6327
No log 2.5455 84 2.0168 -0.0500 2.0168 1.4202
No log 2.6061 86 1.4866 -0.0247 1.4866 1.2193
No log 2.6667 88 1.2210 -0.0466 1.2210 1.1050
No log 2.7273 90 1.0657 -0.1538 1.0657 1.0324
No log 2.7879 92 0.9738 -0.1245 0.9738 0.9868
No log 2.8485 94 0.8630 -0.0031 0.8630 0.9290
No log 2.9091 96 0.8562 -0.0008 0.8562 0.9253
No log 2.9697 98 0.7471 0.0355 0.7471 0.8643
No log 3.0303 100 0.8751 0.1064 0.8751 0.9354
No log 3.0909 102 1.0165 -0.0200 1.0165 1.0082
No log 3.1515 104 0.7625 0.0282 0.7625 0.8732
No log 3.2121 106 0.7475 0.0479 0.7475 0.8646
No log 3.2727 108 0.7740 -0.0967 0.7740 0.8798
No log 3.3333 110 0.7814 0.0395 0.7814 0.8839
No log 3.3939 112 0.8048 0.0318 0.8048 0.8971
No log 3.4545 114 1.3280 -0.0319 1.3280 1.1524
No log 3.5152 116 1.7283 -0.0422 1.7283 1.3147
No log 3.5758 118 1.1934 -0.0575 1.1934 1.0924
No log 3.6364 120 0.7599 -0.0033 0.7599 0.8717
No log 3.6970 122 0.7708 -0.1067 0.7708 0.8779
No log 3.7576 124 0.7589 -0.0571 0.7589 0.8711
No log 3.8182 126 1.0157 -0.0456 1.0157 1.0078
No log 3.8788 128 1.5689 0.0162 1.5689 1.2525
No log 3.9394 130 1.6987 0.0093 1.6987 1.3033
No log 4.0 132 1.2487 -0.0571 1.2487 1.1175
No log 4.0606 134 0.8520 -0.1191 0.8520 0.9230
No log 4.1212 136 0.9168 -0.0376 0.9168 0.9575
No log 4.1818 138 0.8407 -0.1131 0.8407 0.9169
No log 4.2424 140 0.8285 0.0323 0.8285 0.9102
No log 4.3030 142 0.8340 0.0570 0.8340 0.9132
No log 4.3636 144 0.8710 -0.0442 0.8710 0.9333
No log 4.4242 146 0.8228 0.0173 0.8228 0.9071
No log 4.4848 148 0.8122 -0.0091 0.8122 0.9012
No log 4.5455 150 1.0702 -0.0456 1.0702 1.0345
No log 4.6061 152 0.9406 0.0016 0.9406 0.9699
No log 4.6667 154 0.7552 -0.0473 0.7552 0.8691
No log 4.7273 156 0.7800 -0.1001 0.7800 0.8832
No log 4.7879 158 0.7194 0.0 0.7194 0.8482
No log 4.8485 160 0.8910 0.1291 0.8910 0.9439
No log 4.9091 162 1.1826 -0.0269 1.1826 1.0875
No log 4.9697 164 0.9595 0.1107 0.9595 0.9796
No log 5.0303 166 0.7371 0.0759 0.7371 0.8586
No log 5.0909 168 0.7016 -0.0033 0.7016 0.8376
No log 5.1515 170 0.7373 0.0814 0.7373 0.8586
No log 5.2121 172 0.9565 0.1445 0.9565 0.9780
No log 5.2727 174 1.0657 0.1311 1.0657 1.0323
No log 5.3333 176 0.8589 -0.0355 0.8589 0.9268
No log 5.3939 178 0.7575 -0.0541 0.7575 0.8703
No log 5.4545 180 0.8338 -0.0121 0.8338 0.9131
No log 5.5152 182 0.8267 -0.1474 0.8267 0.9092
No log 5.5758 184 0.7688 0.0089 0.7688 0.8768
No log 5.6364 186 0.9301 -0.0861 0.9301 0.9644
No log 5.6970 188 1.1398 -0.0909 1.1398 1.0676
No log 5.7576 190 0.9141 -0.0809 0.9141 0.9561
No log 5.8182 192 0.7624 0.0479 0.7624 0.8731
No log 5.8788 194 0.7702 0.0436 0.7702 0.8776
No log 5.9394 196 0.8365 0.0600 0.8365 0.9146
No log 6.0 198 0.9267 -0.0373 0.9267 0.9626
No log 6.0606 200 0.8381 0.1573 0.8381 0.9155
No log 6.1212 202 0.8592 0.0392 0.8592 0.9269
No log 6.1818 204 0.9043 -0.0289 0.9043 0.9510
No log 6.2424 206 0.8381 -0.1298 0.8381 0.9155
No log 6.3030 208 0.8196 0.0269 0.8196 0.9053
No log 6.3636 210 1.1919 -0.0558 1.1919 1.0918
No log 6.4242 212 1.2191 -0.0571 1.2191 1.1041
No log 6.4848 214 0.9404 -0.0809 0.9404 0.9697
No log 6.5455 216 0.7457 -0.0493 0.7457 0.8635
No log 6.6061 218 0.7302 -0.1001 0.7302 0.8545
No log 6.6667 220 0.7368 -0.0032 0.7368 0.8584
No log 6.7273 222 0.9490 -0.0818 0.9490 0.9741
No log 6.7879 224 0.9726 -0.0456 0.9726 0.9862
No log 6.8485 226 0.7833 0.0732 0.7833 0.8850
No log 6.9091 228 0.7276 0.0094 0.7276 0.8530
No log 6.9697 230 0.7300 -0.1067 0.7300 0.8544
No log 7.0303 232 0.7355 -0.0499 0.7355 0.8576
No log 7.0909 234 0.7109 -0.0551 0.7109 0.8432
No log 7.1515 236 0.7164 -0.0033 0.7164 0.8464
No log 7.2121 238 0.7274 -0.0541 0.7274 0.8529
No log 7.2727 240 0.7227 -0.0551 0.7227 0.8501
No log 7.3333 242 0.7441 -0.0451 0.7441 0.8626
No log 7.3939 244 0.7342 -0.0551 0.7342 0.8569
No log 7.4545 246 0.7725 0.0863 0.7725 0.8789
No log 7.5152 248 0.8116 0.0723 0.8116 0.9009
No log 7.5758 250 0.7492 0.0436 0.7492 0.8656
No log 7.6364 252 0.7762 -0.0551 0.7762 0.8810
No log 7.6970 254 0.8096 -0.0551 0.8096 0.8998
No log 7.7576 256 0.8489 0.0436 0.8489 0.9214
No log 7.8182 258 0.9024 -0.0643 0.9024 0.9499
No log 7.8788 260 0.9086 -0.1271 0.9086 0.9532
No log 7.9394 262 0.9611 -0.0016 0.9611 0.9803
No log 8.0 264 0.9405 -0.0566 0.9405 0.9698
No log 8.0606 266 0.9605 -0.1131 0.9605 0.9800
No log 8.1212 268 1.0400 -0.1214 1.0400 1.0198
No log 8.1818 270 0.9594 -0.1203 0.9594 0.9795
No log 8.2424 272 0.8211 -0.1153 0.8211 0.9062
No log 8.3030 274 0.7834 0.0 0.7834 0.8851
No log 8.3636 276 0.7928 -0.0033 0.7928 0.8904
No log 8.4242 278 0.8399 -0.1180 0.8399 0.9165
No log 8.4848 280 0.8648 -0.0643 0.8648 0.9299
No log 8.5455 282 0.8885 0.0165 0.8885 0.9426
No log 8.6061 284 0.9744 -0.0142 0.9744 0.9871
No log 8.6667 286 0.9912 0.0267 0.9912 0.9956
No log 8.7273 288 0.9116 0.0748 0.9116 0.9548
No log 8.7879 290 0.8904 0.0097 0.8904 0.9436
No log 8.8485 292 1.0948 -0.0151 1.0948 1.0463
No log 8.9091 294 1.5216 -0.1538 1.5216 1.2335
No log 8.9697 296 1.5008 -0.0736 1.5008 1.2251
No log 9.0303 298 1.1821 -0.0686 1.1821 1.0872
No log 9.0909 300 0.8899 -0.0371 0.8899 0.9433
No log 9.1515 302 0.8016 -0.0571 0.8016 0.8953
No log 9.2121 304 0.8192 -0.0551 0.8192 0.9051
No log 9.2727 306 0.8433 0.0064 0.8433 0.9183
No log 9.3333 308 0.8547 0.0030 0.8547 0.9245
No log 9.3939 310 0.8749 -0.1054 0.8749 0.9354
No log 9.4545 312 0.8581 -0.1457 0.8581 0.9263
No log 9.5152 314 0.8135 -0.1088 0.8135 0.9020
No log 9.5758 316 0.7710 -0.1074 0.7710 0.8781
No log 9.6364 318 0.7648 -0.1067 0.7648 0.8745
No log 9.6970 320 0.7779 -0.0704 0.7779 0.8820
No log 9.7576 322 0.9057 -0.0518 0.9057 0.9517
No log 9.8182 324 1.0504 0.0089 1.0504 1.0249
No log 9.8788 326 0.9623 0.0182 0.9623 0.9810
No log 9.9394 328 0.7903 -0.0704 0.7903 0.8890
No log 10.0 330 0.7888 -0.1001 0.7888 0.8881
No log 10.0606 332 0.8225 -0.0881 0.8225 0.9069
No log 10.1212 334 0.8169 -0.0407 0.8169 0.9039
No log 10.1818 336 0.9466 -0.1169 0.9466 0.9729
No log 10.2424 338 1.1011 -0.0885 1.1011 1.0493
No log 10.3030 340 1.0291 -0.0468 1.0291 1.0144
No log 10.3636 342 0.9177 -0.1004 0.9177 0.9579
No log 10.4242 344 0.8824 -0.0187 0.8824 0.9393
No log 10.4848 346 0.8641 0.0187 0.8641 0.9295
No log 10.5455 348 0.8762 -0.1191 0.8762 0.9360
No log 10.6061 350 0.9227 -0.1614 0.9227 0.9606
No log 10.6667 352 0.8844 -0.1200 0.8844 0.9404
No log 10.7273 354 0.9174 -0.0842 0.9174 0.9578
No log 10.7879 356 0.9239 -0.0842 0.9239 0.9612
No log 10.8485 358 0.8806 -0.1197 0.8806 0.9384
No log 10.9091 360 0.9252 -0.1203 0.9252 0.9619
No log 10.9697 362 0.9311 -0.1609 0.9310 0.9649
No log 11.0303 364 0.8672 -0.1180 0.8672 0.9312
No log 11.0909 366 0.8523 -0.2221 0.8523 0.9232
No log 11.1515 368 0.8771 -0.1682 0.8771 0.9366
No log 11.2121 370 0.8877 -0.2221 0.8877 0.9422
No log 11.2727 372 0.9125 -0.0704 0.9125 0.9552
No log 11.3333 374 0.9601 -0.1197 0.9601 0.9799
No log 11.3939 376 0.9411 -0.1191 0.9411 0.9701
No log 11.4545 378 0.9092 -0.2048 0.9092 0.9535
No log 11.5152 380 0.8698 -0.0550 0.8698 0.9326
No log 11.5758 382 0.8597 -0.0595 0.8597 0.9272
No log 11.6364 384 0.8760 -0.2048 0.8760 0.9360
No log 11.6970 386 0.8307 -0.1116 0.8307 0.9114
No log 11.7576 388 0.8546 -0.0718 0.8546 0.9245
No log 11.8182 390 0.8730 -0.0391 0.8730 0.9344
No log 11.8788 392 0.8095 -0.0718 0.8095 0.8997
No log 11.9394 394 0.7864 0.0061 0.7864 0.8868
No log 12.0 396 0.8183 -0.0322 0.8183 0.9046
No log 12.0606 398 0.7993 0.0543 0.7993 0.8940
No log 12.1212 400 0.8260 -0.0252 0.8260 0.9088
No log 12.1818 402 1.0940 -0.0200 1.0940 1.0459
No log 12.2424 404 1.2229 -0.0628 1.2229 1.1058
No log 12.3030 406 1.0898 -0.0143 1.0898 1.0439
No log 12.3636 408 0.8881 -0.0132 0.8881 0.9424
No log 12.4242 410 0.8491 0.0856 0.8491 0.9215
No log 12.4848 412 0.8522 0.0856 0.8522 0.9231
No log 12.5455 414 0.8763 0.0611 0.8763 0.9361
No log 12.6061 416 1.0066 -0.0504 1.0066 1.0033
No log 12.6667 418 1.0041 -0.0854 1.0041 1.0020
No log 12.7273 420 1.0108 -0.0532 1.0108 1.0054
No log 12.7879 422 0.9386 -0.0056 0.9386 0.9688
No log 12.8485 424 0.8531 0.0600 0.8531 0.9236
No log 12.9091 426 0.7970 -0.0163 0.7970 0.8928
No log 12.9697 428 0.8196 0.0600 0.8196 0.9053
No log 13.0303 430 0.9176 -0.0101 0.9176 0.9579
No log 13.0909 432 0.9867 -0.0558 0.9867 0.9933
No log 13.1515 434 0.9229 -0.0902 0.9229 0.9607
No log 13.2121 436 0.7954 0.0768 0.7954 0.8919
No log 13.2727 438 0.7673 -0.0473 0.7673 0.8760
No log 13.3333 440 0.8265 0.0206 0.8265 0.9091
No log 13.3939 442 0.8295 0.0206 0.8295 0.9108
No log 13.4545 444 0.7953 -0.1010 0.7953 0.8918
No log 13.5152 446 0.8357 0.0639 0.8357 0.9142
No log 13.5758 448 0.9867 -0.0545 0.9867 0.9934
No log 13.6364 450 1.0527 -0.0218 1.0527 1.0260
No log 13.6970 452 0.9334 -0.0799 0.9334 0.9661
No log 13.7576 454 0.8252 0.0600 0.8252 0.9084
No log 13.8182 456 0.8057 -0.0252 0.8057 0.8976
No log 13.8788 458 0.8234 0.0600 0.8234 0.9074
No log 13.9394 460 0.7980 -0.0252 0.7980 0.8933
No log 14.0 462 0.7807 -0.0675 0.7807 0.8836
No log 14.0606 464 0.7972 -0.0204 0.7972 0.8928
No log 14.1212 466 0.7983 -0.0675 0.7983 0.8935
No log 14.1818 468 0.7940 -0.0675 0.7940 0.8911
No log 14.2424 470 0.8119 -0.0252 0.8119 0.9011
No log 14.3030 472 0.8225 0.0183 0.8225 0.9069
No log 14.3636 474 0.8490 0.0600 0.8490 0.9214
No log 14.4242 476 0.8564 0.0999 0.8564 0.9254
No log 14.4848 478 0.8263 0.0183 0.8263 0.9090
No log 14.5455 480 0.8254 0.0116 0.8254 0.9085
No log 14.6061 482 0.8469 -0.0322 0.8469 0.9203
No log 14.6667 484 0.8277 0.0543 0.8277 0.9098
No log 14.7273 486 0.8129 0.0454 0.8129 0.9016
No log 14.7879 488 0.8066 0.0863 0.8066 0.8981
No log 14.8485 490 0.8522 -0.0425 0.8522 0.9232
No log 14.9091 492 0.9841 -0.0236 0.9841 0.9920
No log 14.9697 494 1.0155 -0.0571 1.0155 1.0077
No log 15.0303 496 0.9613 -0.0425 0.9613 0.9805
No log 15.0909 498 0.9267 -0.0425 0.9267 0.9627
0.3542 15.1515 500 0.8408 -0.0316 0.8408 0.9170
0.3542 15.2121 502 0.8126 -0.0091 0.8126 0.9014
0.3542 15.2727 504 0.8324 -0.0612 0.8324 0.9124
0.3542 15.3333 506 0.8384 0.0269 0.8384 0.9157
0.3542 15.3939 508 0.8822 -0.1200 0.8822 0.9392
0.3542 15.4545 510 0.8642 -0.1200 0.8642 0.9296
0.3542 15.5152 512 0.8575 0.0289 0.8575 0.9260
0.3542 15.5758 514 0.8868 -0.0025 0.8868 0.9417
0.3542 15.6364 516 0.9098 0.0166 0.9098 0.9538
0.3542 15.6970 518 0.9092 0.0623 0.9092 0.9535
0.3542 15.7576 520 0.8747 0.0940 0.8747 0.9352
0.3542 15.8182 522 0.8626 -0.1180 0.8627 0.9288
0.3542 15.8788 524 0.8997 -0.0390 0.8997 0.9485
0.3542 15.9394 526 0.9250 0.0287 0.9250 0.9618
0.3542 16.0 528 0.8859 -0.0079 0.8859 0.9412
0.3542 16.0606 530 0.8075 -0.1244 0.8075 0.8986
0.3542 16.1212 532 0.7756 -0.0033 0.7756 0.8807
0.3542 16.1818 534 0.7706 -0.0033 0.7706 0.8778
0.3542 16.2424 536 0.7884 -0.0704 0.7884 0.8879
0.3542 16.3030 538 0.8012 -0.0753 0.8012 0.8951
0.3542 16.3636 540 0.8239 -0.0331 0.8239 0.9077
0.3542 16.4242 542 0.8397 -0.0731 0.8397 0.9164
0.3542 16.4848 544 0.8513 -0.0690 0.8513 0.9227

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k13_task3_organization

Finetuned
(4019)
this model