ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k4_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0028
  • Qwk: -0.0595
  • Mse: 1.0028
  • Rmse: 1.0014

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 4.0906 -0.0089 4.0906 2.0225
No log 0.4 4 2.4420 0.0516 2.4420 1.5627
No log 0.6 6 1.6853 0.0504 1.6853 1.2982
No log 0.8 8 1.5679 0.0014 1.5679 1.2521
No log 1.0 10 2.0685 -0.0265 2.0685 1.4382
No log 1.2 12 2.3664 -0.0152 2.3664 1.5383
No log 1.4 14 1.0052 -0.0583 1.0052 1.0026
No log 1.6 16 0.7669 0.1506 0.7669 0.8758
No log 1.8 18 0.9946 -0.0987 0.9946 0.9973
No log 2.0 20 1.0007 -0.0728 1.0007 1.0003
No log 2.2 22 0.9556 -0.0997 0.9556 0.9776
No log 2.4 24 0.9215 0.0067 0.9215 0.9600
No log 2.6 26 1.0086 -0.0117 1.0086 1.0043
No log 2.8 28 1.0934 -0.0712 1.0934 1.0456
No log 3.0 30 1.0786 0.0238 1.0786 1.0385
No log 3.2 32 1.1433 0.0274 1.1433 1.0693
No log 3.4 34 1.2787 -0.0029 1.2787 1.1308
No log 3.6 36 1.0347 -0.0551 1.0347 1.0172
No log 3.8 38 0.8315 -0.0578 0.8315 0.9119
No log 4.0 40 0.8287 -0.0578 0.8287 0.9103
No log 4.2 42 1.0672 -0.0194 1.0672 1.0330
No log 4.4 44 1.6120 -0.0700 1.6120 1.2696
No log 4.6 46 1.3909 -0.0098 1.3909 1.1794
No log 4.8 48 0.9381 -0.2354 0.9381 0.9686
No log 5.0 50 0.8432 0.0122 0.8432 0.9182
No log 5.2 52 0.8024 0.0 0.8024 0.8958
No log 5.4 54 0.8357 -0.1172 0.8357 0.9141
No log 5.6 56 1.3005 -0.0736 1.3005 1.1404
No log 5.8 58 1.7061 0.0235 1.7061 1.3062
No log 6.0 60 1.6126 0.0235 1.6126 1.2699
No log 6.2 62 1.1297 -0.1015 1.1297 1.0629
No log 6.4 64 0.8180 0.0476 0.8180 0.9044
No log 6.6 66 0.7472 0.0033 0.7472 0.8644
No log 6.8 68 0.7974 0.0970 0.7974 0.8930
No log 7.0 70 1.1697 -0.0877 1.1697 1.0815
No log 7.2 72 1.4428 0.0512 1.4428 1.2012
No log 7.4 74 1.1127 -0.1230 1.1127 1.0548
No log 7.6 76 0.8319 0.0028 0.8319 0.9121
No log 7.8 78 0.8058 0.0058 0.8058 0.8977
No log 8.0 80 0.7938 0.0714 0.7938 0.8910
No log 8.2 82 0.8699 -0.0425 0.8699 0.9327
No log 8.4 84 0.8787 -0.0442 0.8787 0.9374
No log 8.6 86 0.8260 0.0670 0.8260 0.9089
No log 8.8 88 0.7770 -0.1010 0.7770 0.8815
No log 9.0 90 0.8042 -0.0091 0.8042 0.8968
No log 9.2 92 1.1137 -0.0182 1.1137 1.0553
No log 9.4 94 1.1995 -0.0923 1.1995 1.0952
No log 9.6 96 0.8840 -0.0371 0.8840 0.9402
No log 9.8 98 0.8149 -0.0532 0.8149 0.9027
No log 10.0 100 0.8142 0.0031 0.8142 0.9023
No log 10.2 102 0.8348 -0.0082 0.8348 0.9137
No log 10.4 104 1.1992 -0.0261 1.1992 1.0951
No log 10.6 106 1.4509 0.0126 1.4509 1.2045
No log 10.8 108 1.0911 -0.0923 1.0911 1.0446
No log 11.0 110 0.7895 -0.0059 0.7895 0.8885
No log 11.2 112 0.7847 0.0031 0.7847 0.8858
No log 11.4 114 0.8415 -0.0113 0.8415 0.9173
No log 11.6 116 1.1146 -0.0840 1.1146 1.0558
No log 11.8 118 1.2182 -0.0905 1.2182 1.1037
No log 12.0 120 0.9284 -0.1088 0.9284 0.9635
No log 12.2 122 0.8088 0.0488 0.8088 0.8993
No log 12.4 124 0.7853 0.0488 0.7853 0.8862
No log 12.6 126 0.8827 -0.0474 0.8827 0.9395
No log 12.8 128 1.0283 0.0089 1.0283 1.0140
No log 13.0 130 0.9379 -0.0143 0.9379 0.9684
No log 13.2 132 0.8496 0.0442 0.8496 0.9217
No log 13.4 134 0.8038 0.0376 0.8038 0.8965
No log 13.6 136 0.8253 0.0528 0.8253 0.9085
No log 13.8 138 0.8280 0.0488 0.8280 0.9100
No log 14.0 140 0.8539 -0.0629 0.8539 0.9241
No log 14.2 142 1.0209 -0.0143 1.0209 1.0104
No log 14.4 144 0.9983 -0.0143 0.9983 0.9991
No log 14.6 146 0.8190 -0.0629 0.8190 0.9050
No log 14.8 148 0.7993 0.0528 0.7993 0.8941
No log 15.0 150 0.8114 0.0926 0.8114 0.9008
No log 15.2 152 0.8917 -0.0376 0.8917 0.9443
No log 15.4 154 1.0242 0.0152 1.0242 1.0120
No log 15.6 156 1.0447 -0.0228 1.0447 1.0221
No log 15.8 158 0.9998 -0.0558 0.9998 0.9999
No log 16.0 160 0.8588 0.0065 0.8588 0.9267
No log 16.2 162 0.7532 0.0471 0.7532 0.8679
No log 16.4 164 0.7291 0.0970 0.7291 0.8539
No log 16.6 166 0.7761 0.0549 0.7761 0.8809
No log 16.8 168 1.0028 0.0046 1.0028 1.0014
No log 17.0 170 1.0804 0.0006 1.0804 1.0394
No log 17.2 172 0.8750 -0.0182 0.8750 0.9354
No log 17.4 174 0.7642 0.1538 0.7642 0.8742
No log 17.6 176 0.8129 -0.0449 0.8129 0.9016
No log 17.8 178 0.8188 0.0571 0.8188 0.9048
No log 18.0 180 0.8899 -0.0616 0.8899 0.9434
No log 18.2 182 0.9261 0.0517 0.9261 0.9624
No log 18.4 184 0.8680 -0.0178 0.8680 0.9317
No log 18.6 186 0.7956 0.0303 0.7956 0.8920
No log 18.8 188 0.7583 0.0303 0.7583 0.8708
No log 19.0 190 0.8025 0.0909 0.8025 0.8958
No log 19.2 192 0.7944 0.0909 0.7944 0.8913
No log 19.4 194 0.7532 0.1318 0.7532 0.8679
No log 19.6 196 0.7700 0.0918 0.7700 0.8775
No log 19.8 198 0.8578 0.0424 0.8578 0.9262
No log 20.0 200 1.0868 -0.0943 1.0868 1.0425
No log 20.2 202 1.3252 -0.0712 1.3252 1.1512
No log 20.4 204 1.3156 -0.0712 1.3156 1.1470
No log 20.6 206 1.1122 -0.1281 1.1122 1.0546
No log 20.8 208 0.8772 0.0650 0.8772 0.9366
No log 21.0 210 0.8057 0.0918 0.8057 0.8976
No log 21.2 212 0.8046 0.0454 0.8046 0.8970
No log 21.4 214 0.8444 0.0588 0.8444 0.9189
No log 21.6 216 0.9596 -0.0909 0.9596 0.9796
No log 21.8 218 1.0491 -0.1278 1.0491 1.0243
No log 22.0 220 0.9715 -0.1270 0.9715 0.9857
No log 22.2 222 0.8996 -0.0156 0.8996 0.9485
No log 22.4 224 0.9159 0.0026 0.9159 0.9570
No log 22.6 226 0.9084 -0.0079 0.9084 0.9531
No log 22.8 228 0.9193 -0.0656 0.9193 0.9588
No log 23.0 230 0.9143 -0.0341 0.9143 0.9562
No log 23.2 232 0.8808 -0.0283 0.8808 0.9385
No log 23.4 234 0.8462 -0.0218 0.8462 0.9199
No log 23.6 236 0.8470 -0.0262 0.8470 0.9203
No log 23.8 238 0.8209 -0.0218 0.8209 0.9061
No log 24.0 240 0.8336 0.0512 0.8336 0.9130
No log 24.2 242 0.9146 -0.0122 0.9146 0.9563
No log 24.4 244 0.9275 -0.0143 0.9275 0.9630
No log 24.6 246 0.8763 -0.0442 0.8763 0.9361
No log 24.8 248 0.8252 0.0611 0.8252 0.9084
No log 25.0 250 0.8338 -0.0218 0.8338 0.9131
No log 25.2 252 0.8439 -0.0614 0.8439 0.9186
No log 25.4 254 0.8325 -0.1060 0.8325 0.9124
No log 25.6 256 0.8269 -0.0583 0.8269 0.9093
No log 25.8 258 0.8647 -0.0283 0.8647 0.9299
No log 26.0 260 0.9353 -0.0490 0.9353 0.9671
No log 26.2 262 0.9384 -0.0518 0.9384 0.9687
No log 26.4 264 0.9028 -0.0490 0.9028 0.9501
No log 26.6 266 0.8395 0.0129 0.8395 0.9163
No log 26.8 268 0.8212 -0.0030 0.8212 0.9062
No log 27.0 270 0.8301 -0.0030 0.8301 0.9111
No log 27.2 272 0.8395 -0.0599 0.8395 0.9163
No log 27.4 274 0.8769 -0.0240 0.8769 0.9364
No log 27.6 276 0.9622 -0.0490 0.9622 0.9809
No log 27.8 278 0.9671 -0.0504 0.9671 0.9834
No log 28.0 280 0.8967 -0.0778 0.8967 0.9470
No log 28.2 282 0.8763 -0.0643 0.8763 0.9361
No log 28.4 284 0.8784 -0.0643 0.8784 0.9372
No log 28.6 286 0.9169 -0.1155 0.9169 0.9575
No log 28.8 288 1.0159 -0.0923 1.0159 1.0079
No log 29.0 290 1.0372 -0.1278 1.0372 1.0184
No log 29.2 292 0.9520 -0.0909 0.9520 0.9757
No log 29.4 294 0.8810 -0.0441 0.8810 0.9386
No log 29.6 296 0.8598 0.1585 0.8598 0.9272
No log 29.8 298 0.8571 0.0289 0.8571 0.9258
No log 30.0 300 0.8905 -0.0746 0.8905 0.9437
No log 30.2 302 0.9159 -0.0490 0.9159 0.9570
No log 30.4 304 0.9372 -0.0532 0.9372 0.9681
No log 30.6 306 0.9377 -0.0518 0.9377 0.9683
No log 30.8 308 0.9105 -0.1557 0.9105 0.9542
No log 31.0 310 0.8980 -0.0200 0.8980 0.9476
No log 31.2 312 0.8922 -0.0271 0.8922 0.9446
No log 31.4 314 0.8772 -0.0723 0.8772 0.9366
No log 31.6 316 0.9054 -0.0459 0.9054 0.9515
No log 31.8 318 0.9272 -0.0518 0.9272 0.9629
No log 32.0 320 0.9017 -0.0459 0.9017 0.9496
No log 32.2 322 0.8635 0.0095 0.8635 0.9292
No log 32.4 324 0.8465 -0.0240 0.8465 0.9201
No log 32.6 326 0.8560 0.0175 0.8560 0.9252
No log 32.8 328 0.9239 -0.0532 0.9239 0.9612
No log 33.0 330 0.9477 -0.0916 0.9477 0.9735
No log 33.2 332 0.9430 -0.0532 0.9430 0.9711
No log 33.4 334 0.8870 -0.0408 0.8870 0.9418
No log 33.6 336 0.8783 -0.0391 0.8783 0.9372
No log 33.8 338 0.9070 -0.0056 0.9070 0.9523
No log 34.0 340 0.9064 -0.0056 0.9064 0.9521
No log 34.2 342 0.9069 -0.0504 0.9069 0.9523
No log 34.4 344 0.8988 -0.0532 0.8988 0.9480
No log 34.6 346 0.8768 -0.0504 0.8768 0.9364
No log 34.8 348 0.8489 0.0043 0.8489 0.9214
No log 35.0 350 0.8055 0.1259 0.8055 0.8975
No log 35.2 352 0.7985 0.0318 0.7985 0.8936
No log 35.4 354 0.8218 0.1202 0.8218 0.9065
No log 35.6 356 0.9251 -0.0474 0.9251 0.9618
No log 35.8 358 1.0204 -0.0923 1.0204 1.0101
No log 36.0 360 1.0196 -0.0200 1.0196 1.0097
No log 36.2 362 0.9281 -0.0504 0.9281 0.9634
No log 36.4 364 0.8590 -0.0390 0.8590 0.9268
No log 36.6 366 0.8070 0.0214 0.8070 0.8983
No log 36.8 368 0.7855 0.1371 0.7855 0.8863
No log 37.0 370 0.8011 0.0214 0.8011 0.8950
No log 37.2 372 0.8539 -0.0425 0.8539 0.9241
No log 37.4 374 0.8835 -0.0474 0.8835 0.9399
No log 37.6 376 0.9485 -0.0163 0.9485 0.9739
No log 37.8 378 1.0164 -0.0606 1.0164 1.0082
No log 38.0 380 1.0498 -0.1282 1.0498 1.0246
No log 38.2 382 1.0279 -0.0253 1.0279 1.0138
No log 38.4 384 0.9476 -0.0490 0.9476 0.9734
No log 38.6 386 0.8772 -0.0008 0.8772 0.9366
No log 38.8 388 0.8552 -0.0371 0.8552 0.9248
No log 39.0 390 0.8818 -0.0474 0.8818 0.9390
No log 39.2 392 0.8945 -0.0504 0.8945 0.9458
No log 39.4 394 0.9166 -0.0902 0.9166 0.9574
No log 39.6 396 0.8893 -0.0504 0.8893 0.9430
No log 39.8 398 0.8704 -0.0459 0.8704 0.9330
No log 40.0 400 0.8766 -0.0459 0.8766 0.9363
No log 40.2 402 0.8877 -0.0459 0.8877 0.9422
No log 40.4 404 0.8819 -0.0425 0.8819 0.9391
No log 40.6 406 0.9101 -0.0474 0.9101 0.9540
No log 40.8 408 0.8853 -0.0425 0.8853 0.9409
No log 41.0 410 0.8720 -0.0425 0.8720 0.9338
No log 41.2 412 0.8917 -0.0459 0.8917 0.9443
No log 41.4 414 0.9177 -0.0474 0.9177 0.9580
No log 41.6 416 0.9121 -0.0101 0.9121 0.9550
No log 41.8 418 0.8782 -0.0442 0.8782 0.9371
No log 42.0 420 0.8605 -0.0331 0.8605 0.9276
No log 42.2 422 0.8759 -0.0459 0.8759 0.9359
No log 42.4 424 0.8756 -0.0459 0.8756 0.9357
No log 42.6 426 0.8546 0.0099 0.8546 0.9245
No log 42.8 428 0.8629 0.0071 0.8629 0.9289
No log 43.0 430 0.8923 -0.0474 0.8923 0.9446
No log 43.2 432 0.9177 -0.0143 0.9177 0.9580
No log 43.4 434 0.9933 -0.1620 0.9933 0.9967
No log 43.6 436 1.0350 -0.0977 1.0350 1.0173
No log 43.8 438 1.0515 -0.0977 1.0515 1.0254
No log 44.0 440 1.0138 -0.1620 1.0138 1.0069
No log 44.2 442 0.9575 -0.0532 0.9575 0.9785
No log 44.4 444 0.9256 -0.0143 0.9256 0.9621
No log 44.6 446 0.9260 -0.0143 0.9260 0.9623
No log 44.8 448 0.9653 -0.0532 0.9653 0.9825
No log 45.0 450 1.0163 -0.1624 1.0163 1.0081
No log 45.2 452 1.0773 -0.0992 1.0773 1.0380
No log 45.4 454 1.0595 -0.0992 1.0595 1.0293
No log 45.6 456 0.9545 -0.1632 0.9545 0.9770
No log 45.8 458 0.8354 -0.0699 0.8354 0.9140
No log 46.0 460 0.8167 0.0918 0.8167 0.9037
No log 46.2 462 0.8032 0.0970 0.8032 0.8962
No log 46.4 464 0.8229 -0.0643 0.8229 0.9071
No log 46.6 466 0.8425 0.0068 0.8425 0.9179
No log 46.8 468 0.8841 -0.0033 0.8841 0.9403
No log 47.0 470 0.9045 -0.0490 0.9045 0.9511
No log 47.2 472 0.8771 -0.0033 0.8771 0.9365
No log 47.4 474 0.8328 0.0068 0.8328 0.9126
No log 47.6 476 0.8224 -0.0252 0.8224 0.9069
No log 47.8 478 0.8298 0.0095 0.8298 0.9109
No log 48.0 480 0.8663 -0.0056 0.8663 0.9307
No log 48.2 482 0.9339 -0.0558 0.9339 0.9664
No log 48.4 484 0.9505 -0.1281 0.9505 0.9749
No log 48.6 486 0.9262 -0.0583 0.9262 0.9624
No log 48.8 488 0.8916 -0.0163 0.8916 0.9443
No log 49.0 490 0.8437 -0.0033 0.8437 0.9185
No log 49.2 492 0.8276 0.0043 0.8276 0.9097
No log 49.4 494 0.8497 -0.0425 0.8497 0.9218
No log 49.6 496 0.9038 0.0287 0.9038 0.9507
No log 49.8 498 0.9530 -0.0163 0.9530 0.9762
0.2163 50.0 500 0.9776 -0.0163 0.9776 0.9887
0.2163 50.2 502 0.9714 -0.0163 0.9714 0.9856
0.2163 50.4 504 0.9422 -0.0163 0.9422 0.9707
0.2163 50.6 506 0.9353 -0.0163 0.9353 0.9671
0.2163 50.8 508 0.9014 0.0287 0.9014 0.9494
0.2163 51.0 510 0.8676 -0.0079 0.8676 0.9315
0.2163 51.2 512 0.8412 -0.0390 0.8412 0.9172
0.2163 51.4 514 0.8317 -0.0390 0.8317 0.9120
0.2163 51.6 516 0.8603 -0.0459 0.8603 0.9275
0.2163 51.8 518 0.8939 0.0287 0.8939 0.9454
0.2163 52.0 520 0.9515 -0.0916 0.9515 0.9755
0.2163 52.2 522 0.9928 -0.0936 0.9928 0.9964
0.2163 52.4 524 0.9951 -0.0936 0.9951 0.9976
0.2163 52.6 526 0.9448 -0.0182 0.9448 0.9720
0.2163 52.8 528 0.9050 0.0260 0.9050 0.9513
0.2163 53.0 530 0.8878 0.0287 0.8878 0.9422
0.2163 53.2 532 0.8714 0.0287 0.8714 0.9335
0.2163 53.4 534 0.8806 0.0287 0.8806 0.9384
0.2163 53.6 536 0.8581 0.0409 0.8581 0.9264
0.2163 53.8 538 0.8794 -0.0079 0.8794 0.9377
0.2163 54.0 540 0.9232 -0.0163 0.9232 0.9608
0.2163 54.2 542 0.9816 -0.0583 0.9816 0.9907
0.2163 54.4 544 1.0520 -0.0695 1.0520 1.0257
0.2163 54.6 546 1.0506 -0.0987 1.0506 1.0250
0.2163 54.8 548 1.0028 -0.0595 1.0028 1.0014

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k4_task3_organization

Finetuned
(4019)
this model