ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k18_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7841
  • Qwk: -0.1917
  • Mse: 0.7841
  • Rmse: 0.8855

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0444 2 3.7194 -0.0157 3.7194 1.9286
No log 0.0889 4 2.3453 0.0388 2.3453 1.5314
No log 0.1333 6 2.3136 0.0504 2.3136 1.5210
No log 0.1778 8 1.6521 0.0 1.6521 1.2853
No log 0.2222 10 0.9837 0.0006 0.9837 0.9918
No log 0.2667 12 0.8564 0.0017 0.8564 0.9254
No log 0.3111 14 0.7882 -0.0264 0.7882 0.8878
No log 0.3556 16 0.8003 -0.0264 0.8003 0.8946
No log 0.4 18 0.8126 -0.0264 0.8126 0.9014
No log 0.4444 20 1.1233 0.0217 1.1233 1.0599
No log 0.4889 22 1.4059 -0.0133 1.4059 1.1857
No log 0.5333 24 1.1049 -0.0013 1.1049 1.0511
No log 0.5778 26 0.8013 -0.1470 0.8013 0.8951
No log 0.6222 28 0.7922 -0.1001 0.7922 0.8901
No log 0.6667 30 0.8672 -0.1633 0.8672 0.9312
No log 0.7111 32 1.0519 0.0217 1.0519 1.0256
No log 0.7556 34 1.4209 -0.0490 1.4209 1.1920
No log 0.8 36 1.1628 -0.1011 1.1628 1.0783
No log 0.8444 38 1.0332 0.0217 1.0332 1.0165
No log 0.8889 40 0.9067 -0.0474 0.9067 0.9522
No log 0.9333 42 0.7844 -0.0551 0.7844 0.8857
No log 0.9778 44 0.7996 -0.0644 0.7996 0.8942
No log 1.0222 46 0.8820 -0.1628 0.8820 0.9391
No log 1.0667 48 0.8148 -0.0228 0.8148 0.9027
No log 1.1111 50 0.8178 -0.1033 0.8178 0.9043
No log 1.1556 52 0.8776 -0.0599 0.8776 0.9368
No log 1.2 54 1.0582 -0.0341 1.0582 1.0287
No log 1.2444 56 0.9704 -0.0629 0.9704 0.9851
No log 1.2889 58 0.9554 -0.1697 0.9554 0.9774
No log 1.3333 60 0.9922 -0.2100 0.9922 0.9961
No log 1.3778 62 1.0022 -0.2175 1.0022 1.0011
No log 1.4222 64 1.0345 -0.2236 1.0345 1.0171
No log 1.4667 66 1.0173 -0.1851 1.0173 1.0086
No log 1.5111 68 1.0074 -0.2741 1.0074 1.0037
No log 1.5556 70 0.9272 -0.2777 0.9272 0.9629
No log 1.6 72 0.9078 -0.2315 0.9078 0.9528
No log 1.6444 74 0.8941 -0.1100 0.8941 0.9456
No log 1.6889 76 1.2321 -0.1616 1.2321 1.1100
No log 1.7333 78 1.0285 -0.1162 1.0285 1.0141
No log 1.7778 80 0.8974 -0.1671 0.8974 0.9473
No log 1.8222 82 0.9052 -0.1871 0.9052 0.9514
No log 1.8667 84 1.0307 0.0392 1.0307 1.0153
No log 1.9111 86 0.8781 -0.1026 0.8781 0.9371
No log 1.9556 88 0.8672 -0.1833 0.8672 0.9313
No log 2.0 90 0.8799 -0.2315 0.8799 0.9381
No log 2.0444 92 0.9216 -0.0595 0.9216 0.9600
No log 2.0889 94 0.9895 -0.0686 0.9895 0.9947
No log 2.1333 96 0.9515 -0.1786 0.9515 0.9754
No log 2.1778 98 0.9749 -0.1715 0.9749 0.9874
No log 2.2222 100 1.0324 -0.1149 1.0324 1.0161
No log 2.2667 102 1.0325 -0.1103 1.0325 1.0161
No log 2.3111 104 1.0174 -0.1786 1.0174 1.0087
No log 2.3556 106 1.0479 -0.1055 1.0479 1.0237
No log 2.4 108 0.8487 -0.2808 0.8487 0.9213
No log 2.4444 110 0.8242 0.0715 0.8242 0.9079
No log 2.4889 112 0.8272 0.0271 0.8272 0.9095
No log 2.5333 114 0.8298 -0.1531 0.8298 0.9109
No log 2.5778 116 0.8400 -0.1111 0.8400 0.9165
No log 2.6222 118 0.8173 -0.1067 0.8173 0.9040
No log 2.6667 120 0.8867 -0.0150 0.8867 0.9416
No log 2.7111 122 0.8325 -0.0284 0.8325 0.9124
No log 2.7556 124 0.7633 -0.1001 0.7633 0.8737
No log 2.8 126 0.8644 0.0214 0.8644 0.9297
No log 2.8444 128 1.1125 -0.0972 1.1125 1.0547
No log 2.8889 130 0.9986 -0.0595 0.9986 0.9993
No log 2.9333 132 0.8070 -0.0179 0.8070 0.8983
No log 2.9778 134 0.7416 -0.0499 0.7416 0.8612
No log 3.0222 136 0.7577 -0.1001 0.7577 0.8705
No log 3.0667 138 0.8358 -0.0252 0.8358 0.9142
No log 3.1111 140 1.3993 -0.0969 1.3993 1.1829
No log 3.1556 142 1.4724 -0.0969 1.4724 1.2134
No log 3.2 144 0.9683 0.0676 0.9683 0.9840
No log 3.2444 146 0.8477 -0.1001 0.8477 0.9207
No log 3.2889 148 0.8996 0.0155 0.8996 0.9485
No log 3.3333 150 0.8385 -0.1001 0.8385 0.9157
No log 3.3778 152 0.8055 -0.1628 0.8055 0.8975
No log 3.4222 154 1.0214 0.0545 1.0214 1.0107
No log 3.4667 156 1.1279 0.0046 1.1279 1.0620
No log 3.5111 158 0.9198 0.0016 0.9198 0.9590
No log 3.5556 160 0.7874 -0.0179 0.7874 0.8874
No log 3.6 162 0.7757 -0.1001 0.7757 0.8808
No log 3.6444 164 0.8187 -0.1001 0.8187 0.9048
No log 3.6889 166 0.8639 -0.1001 0.8639 0.9295
No log 3.7333 168 0.8406 -0.0091 0.8406 0.9169
No log 3.7778 170 0.8803 -0.0218 0.8803 0.9382
No log 3.8222 172 0.8547 -0.0658 0.8547 0.9245
No log 3.8667 174 0.8242 -0.1081 0.8242 0.9078
No log 3.9111 176 0.8849 -0.2840 0.8849 0.9407
No log 3.9556 178 0.8394 -0.1001 0.8394 0.9162
No log 4.0 180 0.7459 -0.0551 0.7459 0.8637
No log 4.0444 182 0.7330 0.0 0.7330 0.8561
No log 4.0889 184 0.7546 -0.1001 0.7546 0.8687
No log 4.1333 186 0.7965 -0.1001 0.7965 0.8924
No log 4.1778 188 0.8363 -0.2006 0.8363 0.9145
No log 4.2222 190 0.8449 -0.1397 0.8449 0.9192
No log 4.2667 192 0.8577 -0.3108 0.8577 0.9261
No log 4.3111 194 0.9914 -0.0211 0.9914 0.9957
No log 4.3556 196 1.0330 0.0531 1.0330 1.0163
No log 4.4 198 0.9699 0.0446 0.9699 0.9848
No log 4.4444 200 0.8737 -0.2006 0.8737 0.9347
No log 4.4889 202 0.8542 0.0512 0.8542 0.9242
No log 4.5333 204 1.1038 -0.0558 1.1038 1.0506
No log 4.5778 206 1.0335 -0.0122 1.0335 1.0166
No log 4.6222 208 0.8553 -0.0550 0.8553 0.9248
No log 4.6667 210 0.8641 -0.1001 0.8641 0.9295
No log 4.7111 212 0.9147 -0.1142 0.9147 0.9564
No log 4.7556 214 0.9007 -0.2243 0.9007 0.9491
No log 4.8 216 0.9129 -0.1033 0.9129 0.9555
No log 4.8444 218 0.9085 -0.0550 0.9085 0.9532
No log 4.8889 220 0.8560 -0.1018 0.8560 0.9252
No log 4.9333 222 0.8140 -0.0513 0.8140 0.9022
No log 4.9778 224 0.7931 -0.0032 0.7931 0.8906
No log 5.0222 226 0.7892 -0.0499 0.7892 0.8883
No log 5.0667 228 0.8108 0.0033 0.8108 0.9004
No log 5.1111 230 0.8045 -0.1001 0.8045 0.8969
No log 5.1556 232 0.8221 -0.1473 0.8221 0.9067
No log 5.2 234 0.8677 -0.2243 0.8677 0.9315
No log 5.2444 236 0.9042 -0.1142 0.9042 0.9509
No log 5.2889 238 0.8788 -0.1473 0.8788 0.9375
No log 5.3333 240 0.9634 -0.1155 0.9634 0.9815
No log 5.3778 242 0.9327 -0.0682 0.9327 0.9658
No log 5.4222 244 0.8460 -0.1473 0.8460 0.9198
No log 5.4667 246 0.8437 0.0206 0.8437 0.9185
No log 5.5111 248 0.7663 -0.1473 0.7663 0.8754
No log 5.5556 250 0.7765 0.1512 0.7765 0.8812
No log 5.6 252 0.7746 0.1512 0.7746 0.8801
No log 5.6444 254 0.7529 -0.1001 0.7529 0.8677
No log 5.6889 256 0.7622 -0.1001 0.7622 0.8730
No log 5.7333 258 0.7587 -0.1001 0.7587 0.8711
No log 5.7778 260 0.7848 -0.0493 0.7848 0.8859
No log 5.8222 262 0.8169 -0.0958 0.8169 0.9038
No log 5.8667 264 0.8473 -0.2336 0.8473 0.9205
No log 5.9111 266 0.8493 -0.2336 0.8493 0.9216
No log 5.9556 268 0.8296 -0.0513 0.8296 0.9108
No log 6.0 270 0.8079 -0.0513 0.8079 0.8988
No log 6.0444 272 0.7923 -0.1018 0.7923 0.8901
No log 6.0889 274 0.7928 -0.1001 0.7928 0.8904
No log 6.1333 276 0.8037 -0.1010 0.8037 0.8965
No log 6.1778 278 0.8864 0.0175 0.8864 0.9415
No log 6.2222 280 0.9097 0.0574 0.9097 0.9538
No log 6.2667 282 0.8365 -0.0493 0.8365 0.9146
No log 6.3111 284 0.8224 -0.1010 0.8224 0.9069
No log 6.3556 286 0.8552 -0.0030 0.8552 0.9248
No log 6.4 288 0.8948 0.0700 0.8948 0.9459
No log 6.4444 290 0.9736 -0.0909 0.9736 0.9867
No log 6.4889 292 0.8914 0.0588 0.8914 0.9441
No log 6.5333 294 0.8465 -0.1678 0.8465 0.9201
No log 6.5778 296 0.8700 -0.1033 0.8700 0.9328
No log 6.6222 298 0.7977 -0.1473 0.7977 0.8931
No log 6.6667 300 0.7994 -0.0062 0.7994 0.8941
No log 6.7111 302 0.7939 -0.0062 0.7939 0.8910
No log 6.7556 304 0.7683 -0.1001 0.7683 0.8765
No log 6.8 306 0.7733 -0.1001 0.7733 0.8794
No log 6.8444 308 0.7445 -0.0473 0.7445 0.8629
No log 6.8889 310 0.8147 0.1146 0.8147 0.9026
No log 6.9333 312 0.8037 0.0414 0.8037 0.8965
No log 6.9778 314 0.8107 -0.1905 0.8107 0.9004
No log 7.0222 316 0.8480 -0.1745 0.8480 0.9209
No log 7.0667 318 0.8204 -0.1470 0.8204 0.9057
No log 7.1111 320 0.8089 -0.1470 0.8089 0.8994
No log 7.1556 322 0.8135 -0.1470 0.8135 0.9019
No log 7.2 324 0.8140 -0.1398 0.8140 0.9022
No log 7.2444 326 0.8128 -0.1470 0.8128 0.9016
No log 7.2889 328 0.8191 -0.2336 0.8191 0.9050
No log 7.3333 330 0.7931 -0.0633 0.7931 0.8906
No log 7.3778 332 0.7373 0.0628 0.7373 0.8586
No log 7.4222 334 0.7825 0.0031 0.7825 0.8846
No log 7.4667 336 0.9106 0.0016 0.9106 0.9542
No log 7.5111 338 0.8299 0.1001 0.8299 0.9110
No log 7.5556 340 0.7352 0.0 0.7352 0.8574
No log 7.6 342 0.7553 -0.0428 0.7553 0.8691
No log 7.6444 344 0.8028 -0.0322 0.8028 0.8960
No log 7.6889 346 0.7606 -0.0837 0.7606 0.8721
No log 7.7333 348 0.7635 -0.1398 0.7635 0.8738
No log 7.7778 350 0.7894 -0.1531 0.7894 0.8885
No log 7.8222 352 0.8829 -0.0230 0.8829 0.9396
No log 7.8667 354 0.8112 -0.1468 0.8112 0.9007
No log 7.9111 356 0.8211 -0.0173 0.8211 0.9061
No log 7.9556 358 0.7925 -0.1466 0.7925 0.8902
No log 8.0 360 0.7703 -0.1018 0.7703 0.8777
No log 8.0444 362 0.7620 -0.0473 0.7620 0.8729
No log 8.0889 364 0.7756 -0.1001 0.7756 0.8807
No log 8.1333 366 0.7845 -0.1001 0.7845 0.8857
No log 8.1778 368 0.7526 -0.0520 0.7526 0.8675
No log 8.2222 370 0.7747 -0.0520 0.7747 0.8802
No log 8.2667 372 0.7810 -0.1470 0.7810 0.8838
No log 8.3111 374 0.8120 -0.1470 0.8120 0.9011
No log 8.3556 376 0.8729 0.0310 0.8729 0.9343
No log 8.4 378 0.9685 0.0800 0.9685 0.9841
No log 8.4444 380 1.0233 0.0531 1.0233 1.0116
No log 8.4889 382 0.9531 -0.2531 0.9531 0.9763
No log 8.5333 384 0.8679 -0.2243 0.8679 0.9316
No log 8.5778 386 0.8360 -0.2006 0.8360 0.9143
No log 8.6222 388 0.8532 -0.2006 0.8532 0.9237
No log 8.6667 390 0.8524 -0.2513 0.8524 0.9232
No log 8.7111 392 0.9060 -0.2195 0.9060 0.9518
No log 8.7556 394 0.8872 -0.3072 0.8872 0.9419
No log 8.8 396 0.8917 -0.2190 0.8917 0.9443
No log 8.8444 398 0.9057 -0.1939 0.9057 0.9517
No log 8.8889 400 0.8849 -0.1893 0.8849 0.9407
No log 8.9333 402 0.8821 -0.2315 0.8821 0.9392
No log 8.9778 404 0.8978 -0.1753 0.8978 0.9475
No log 9.0222 406 0.8239 -0.2513 0.8239 0.9077
No log 9.0667 408 0.8441 -0.0125 0.8441 0.9187
No log 9.1111 410 0.8517 -0.1106 0.8517 0.9229
No log 9.1556 412 0.8298 -0.1833 0.8298 0.9109
No log 9.2 414 0.8624 -0.2274 0.8624 0.9286
No log 9.2444 416 0.8776 -0.0887 0.8776 0.9368
No log 9.2889 418 0.8114 -0.0826 0.8114 0.9008
No log 9.3333 420 0.7508 -0.0473 0.7508 0.8665
No log 9.3778 422 0.7294 0.0033 0.7294 0.8541
No log 9.4222 424 0.7019 0.0 0.7019 0.8378
No log 9.4667 426 0.7151 0.1023 0.7151 0.8457
No log 9.5111 428 0.7532 -0.0473 0.7532 0.8679
No log 9.5556 430 0.8322 -0.1354 0.8322 0.9123
No log 9.6 432 0.8630 -0.0733 0.8630 0.9290
No log 9.6444 434 0.8676 -0.0558 0.8676 0.9314
No log 9.6889 436 0.9238 -0.0910 0.9238 0.9612
No log 9.7333 438 0.9920 -0.0486 0.9920 0.9960
No log 9.7778 440 0.9290 -0.0008 0.9290 0.9638
No log 9.8222 442 0.8341 -0.1542 0.8341 0.9133
No log 9.8667 444 0.8198 -0.1329 0.8198 0.9054
No log 9.9111 446 0.7983 -0.1001 0.7983 0.8935
No log 9.9556 448 0.7564 0.0033 0.7564 0.8697
No log 10.0 450 0.8841 0.0871 0.8841 0.9403
No log 10.0444 452 0.9671 -0.1221 0.9671 0.9834
No log 10.0889 454 0.8757 0.0146 0.8757 0.9358
No log 10.1333 456 0.8108 -0.1978 0.8108 0.9005
No log 10.1778 458 0.8168 -0.1398 0.8168 0.9038
No log 10.2222 460 0.7912 -0.1018 0.7912 0.8895
No log 10.2667 462 0.8576 0.0490 0.8576 0.9261
No log 10.3111 464 0.8997 0.0786 0.8997 0.9485
No log 10.3556 466 0.8010 0.1506 0.8010 0.8950
No log 10.4 468 0.7215 0.0555 0.7215 0.8494
No log 10.4444 470 0.8170 -0.1001 0.8170 0.9039
No log 10.4889 472 0.8304 0.0685 0.8304 0.9113
No log 10.5333 474 0.7781 -0.0520 0.7781 0.8821
No log 10.5778 476 0.9371 0.0016 0.9371 0.9680
No log 10.6222 478 1.0690 -0.0961 1.0690 1.0339
No log 10.6667 480 1.0825 -0.1278 1.0825 1.0404
No log 10.7111 482 0.9644 -0.0778 0.9644 0.9821
No log 10.7556 484 0.7827 -0.1081 0.7827 0.8847
No log 10.8 486 0.7967 0.0094 0.7967 0.8926
No log 10.8444 488 0.8143 -0.1018 0.8143 0.9024
No log 10.8889 490 0.9216 -0.1077 0.9216 0.9600
No log 10.9333 492 1.0858 -0.1869 1.0858 1.0420
No log 10.9778 494 1.2750 -0.1830 1.2750 1.1291
No log 11.0222 496 1.3253 -0.1848 1.3253 1.1512
No log 11.0667 498 1.1665 -0.2191 1.1665 1.0801
0.2827 11.1111 500 0.9410 -0.0541 0.9410 0.9701
0.2827 11.1556 502 0.8965 -0.2036 0.8965 0.9468
0.2827 11.2 504 0.8787 -0.1086 0.8787 0.9374
0.2827 11.2444 506 0.7974 -0.1470 0.7974 0.8929
0.2827 11.2889 508 0.8658 0.1001 0.8658 0.9305
0.2827 11.3333 510 1.0303 -0.1278 1.0303 1.0150
0.2827 11.3778 512 1.0191 -0.1279 1.0191 1.0095
0.2827 11.4222 514 0.8498 0.1395 0.8498 0.9219
0.2827 11.4667 516 0.6994 -0.0520 0.6994 0.8363
0.2827 11.5111 518 0.7665 -0.1001 0.7665 0.8755
0.2827 11.5556 520 0.7981 -0.1329 0.7981 0.8933
0.2827 11.6 522 0.7848 -0.1398 0.7848 0.8859
0.2827 11.6444 524 0.8517 0.0828 0.8517 0.9229
0.2827 11.6889 526 0.9200 0.0490 0.9200 0.9592
0.2827 11.7333 528 1.0294 -0.1273 1.0294 1.0146
0.2827 11.7778 530 0.9840 -0.0878 0.9840 0.9920
0.2827 11.8222 532 0.8008 0.0918 0.8008 0.8949
0.2827 11.8667 534 0.7453 -0.0473 0.7453 0.8633
0.2827 11.9111 536 0.7676 -0.1001 0.7676 0.8761
0.2827 11.9556 538 0.7633 -0.0493 0.7633 0.8737
0.2827 12.0 540 0.8012 0.1423 0.8013 0.8951
0.2827 12.0444 542 0.8646 0.0183 0.8646 0.9298
0.2827 12.0889 544 0.7911 0.1423 0.7911 0.8894
0.2827 12.1333 546 0.7568 -0.0493 0.7568 0.8700
0.2827 12.1778 548 0.7924 -0.2006 0.7924 0.8902
0.2827 12.2222 550 0.7726 -0.2006 0.7726 0.8790
0.2827 12.2667 552 0.7643 -0.1547 0.7643 0.8743
0.2827 12.3111 554 0.7871 -0.1542 0.7871 0.8872
0.2827 12.3556 556 0.7841 -0.1917 0.7841 0.8855

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k18_task3_organization

Finetuned
(4019)
this model