ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k6_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0184
  • Qwk: 0.6581
  • Mse: 1.0184
  • Rmse: 1.0092

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0444 2 6.8543 0.0308 6.8543 2.6181
No log 0.0889 4 4.5907 0.0873 4.5907 2.1426
No log 0.1333 6 2.7788 0.0848 2.7788 1.6670
No log 0.1778 8 2.0518 0.1343 2.0518 1.4324
No log 0.2222 10 1.8008 0.2202 1.8008 1.3420
No log 0.2667 12 1.8796 0.1682 1.8796 1.3710
No log 0.3111 14 2.0630 0.1429 2.0630 1.4363
No log 0.3556 16 1.9611 0.1698 1.9611 1.4004
No log 0.4 18 1.9650 0.1667 1.9650 1.4018
No log 0.4444 20 1.8260 0.1905 1.8260 1.3513
No log 0.4889 22 1.6127 0.1714 1.6127 1.2699
No log 0.5333 24 1.6073 0.2727 1.6073 1.2678
No log 0.5778 26 1.5918 0.4098 1.5918 1.2617
No log 0.6222 28 1.5010 0.4127 1.5010 1.2252
No log 0.6667 30 1.3706 0.3697 1.3706 1.1707
No log 0.7111 32 1.3654 0.4444 1.3654 1.1685
No log 0.7556 34 1.5157 0.4341 1.5157 1.2311
No log 0.8 36 1.6958 0.2969 1.6958 1.3022
No log 0.8444 38 1.8635 0.1356 1.8635 1.3651
No log 0.8889 40 1.6984 0.2636 1.6984 1.3032
No log 0.9333 42 1.4409 0.4688 1.4409 1.2004
No log 0.9778 44 1.1620 0.544 1.1620 1.0779
No log 1.0222 46 1.0819 0.5484 1.0819 1.0401
No log 1.0667 48 1.0536 0.5854 1.0536 1.0264
No log 1.1111 50 1.0432 0.4348 1.0432 1.0214
No log 1.1556 52 1.0811 0.4576 1.0811 1.0398
No log 1.2 54 1.2880 0.4202 1.2880 1.1349
No log 1.2444 56 1.3114 0.4333 1.3114 1.1452
No log 1.2889 58 1.3225 0.4590 1.3225 1.1500
No log 1.3333 60 1.1226 0.6047 1.1226 1.0595
No log 1.3778 62 0.8372 0.6715 0.8372 0.9150
No log 1.4222 64 0.8808 0.5736 0.8808 0.9385
No log 1.4667 66 0.8747 0.6763 0.8747 0.9353
No log 1.5111 68 1.2595 0.5185 1.2595 1.1223
No log 1.5556 70 1.4525 0.4444 1.4525 1.2052
No log 1.6 72 1.3101 0.4848 1.3101 1.1446
No log 1.6444 74 1.0378 0.6269 1.0378 1.0187
No log 1.6889 76 0.8729 0.6917 0.8729 0.9343
No log 1.7333 78 0.8617 0.6716 0.8617 0.9283
No log 1.7778 80 1.0152 0.6029 1.0152 1.0075
No log 1.8222 82 1.3002 0.4962 1.3002 1.1403
No log 1.8667 84 1.4579 0.4889 1.4579 1.2074
No log 1.9111 86 1.2173 0.5522 1.2173 1.1033
No log 1.9556 88 0.8831 0.6667 0.8831 0.9397
No log 2.0 90 0.8163 0.6857 0.8163 0.9035
No log 2.0444 92 0.9970 0.6528 0.9970 0.9985
No log 2.0889 94 1.1047 0.5811 1.1047 1.0510
No log 2.1333 96 1.0255 0.6267 1.0255 1.0127
No log 2.1778 98 1.1141 0.6081 1.1141 1.0555
No log 2.2222 100 1.3814 0.5395 1.3814 1.1753
No log 2.2667 102 1.4554 0.4091 1.4554 1.2064
No log 2.3111 104 1.5487 0.2400 1.5487 1.2445
No log 2.3556 106 1.4966 0.3871 1.4966 1.2234
No log 2.4 108 1.1061 0.5203 1.1061 1.0517
No log 2.4444 110 0.9115 0.5984 0.9115 0.9548
No log 2.4889 112 1.2691 0.5156 1.2690 1.1265
No log 2.5333 114 1.3156 0.4516 1.3156 1.1470
No log 2.5778 116 1.0065 0.5410 1.0065 1.0032
No log 2.6222 118 1.0274 0.6308 1.0274 1.0136
No log 2.6667 120 1.2010 0.5441 1.2010 1.0959
No log 2.7111 122 1.2550 0.5255 1.2550 1.1203
No log 2.7556 124 1.1840 0.5481 1.1840 1.0881
No log 2.8 126 0.9876 0.6324 0.9876 0.9938
No log 2.8444 128 0.9159 0.6569 0.9159 0.9570
No log 2.8889 130 0.9374 0.6434 0.9374 0.9682
No log 2.9333 132 1.0694 0.64 1.0694 1.0341
No log 2.9778 134 1.3808 0.5562 1.3808 1.1751
No log 3.0222 136 1.4684 0.5333 1.4684 1.2118
No log 3.0667 138 1.2976 0.5278 1.2976 1.1391
No log 3.1111 140 1.0418 0.6165 1.0418 1.0207
No log 3.1556 142 0.9137 0.6260 0.9137 0.9559
No log 3.2 144 0.8943 0.6418 0.8943 0.9457
No log 3.2444 146 0.9288 0.6269 0.9288 0.9637
No log 3.2889 148 1.0406 0.6074 1.0406 1.0201
No log 3.3333 150 1.3127 0.5714 1.3127 1.1457
No log 3.3778 152 1.4485 0.4967 1.4485 1.2035
No log 3.4222 154 1.3136 0.4861 1.3136 1.1461
No log 3.4667 156 1.1219 0.5672 1.1219 1.0592
No log 3.5111 158 0.9206 0.6212 0.9206 0.9595
No log 3.5556 160 0.8463 0.6316 0.8463 0.9199
No log 3.6 162 0.8714 0.6713 0.8714 0.9335
No log 3.6444 164 0.9831 0.6623 0.9831 0.9915
No log 3.6889 166 1.0459 0.6667 1.0459 1.0227
No log 3.7333 168 0.9525 0.6623 0.9525 0.9760
No log 3.7778 170 0.9800 0.6490 0.9800 0.9899
No log 3.8222 172 0.9415 0.6573 0.9415 0.9703
No log 3.8667 174 0.9812 0.6286 0.9812 0.9906
No log 3.9111 176 1.0709 0.5857 1.0709 1.0348
No log 3.9556 178 1.0794 0.6014 1.0794 1.0389
No log 4.0 180 1.1048 0.5839 1.1048 1.0511
No log 4.0444 182 1.1009 0.6014 1.1009 1.0492
No log 4.0889 184 0.9828 0.5957 0.9828 0.9914
No log 4.1333 186 0.9885 0.6111 0.9885 0.9942
No log 4.1778 188 1.0383 0.6358 1.0383 1.0190
No log 4.2222 190 0.9396 0.6892 0.9396 0.9693
No log 4.2667 192 0.8734 0.6759 0.8734 0.9346
No log 4.3111 194 0.7623 0.7162 0.7623 0.8731
No log 4.3556 196 0.8256 0.7075 0.8256 0.9086
No log 4.4 198 1.0753 0.6282 1.0753 1.0370
No log 4.4444 200 1.0541 0.6405 1.0541 1.0267
No log 4.4889 202 0.8831 0.6619 0.8831 0.9397
No log 4.5333 204 0.7570 0.6815 0.7570 0.8701
No log 4.5778 206 0.7610 0.6716 0.7610 0.8724
No log 4.6222 208 0.8748 0.6713 0.8748 0.9353
No log 4.6667 210 1.0145 0.6275 1.0145 1.0072
No log 4.7111 212 0.9874 0.6309 0.9874 0.9937
No log 4.7556 214 0.8753 0.6892 0.8753 0.9356
No log 4.8 216 0.8755 0.6755 0.8755 0.9357
No log 4.8444 218 0.8946 0.6456 0.8946 0.9458
No log 4.8889 220 0.8821 0.6752 0.8821 0.9392
No log 4.9333 222 0.8233 0.7105 0.8233 0.9074
No log 4.9778 224 0.7758 0.7162 0.7758 0.8808
No log 5.0222 226 0.7170 0.7194 0.7170 0.8468
No log 5.0667 228 0.7129 0.7 0.7129 0.8443
No log 5.1111 230 0.7539 0.7114 0.7539 0.8683
No log 5.1556 232 0.9762 0.6667 0.9762 0.9880
No log 5.2 234 0.9591 0.6918 0.9591 0.9793
No log 5.2444 236 0.8153 0.72 0.8153 0.9029
No log 5.2889 238 0.7302 0.72 0.7302 0.8545
No log 5.3333 240 0.7220 0.7007 0.7220 0.8497
No log 5.3778 242 0.8271 0.6667 0.8271 0.9095
No log 5.4222 244 0.9721 0.6800 0.9721 0.9859
No log 5.4667 246 0.9501 0.7027 0.9501 0.9748
No log 5.5111 248 0.8430 0.6667 0.8430 0.9181
No log 5.5556 250 0.7479 0.6912 0.7479 0.8648
No log 5.6 252 0.7280 0.7068 0.7280 0.8532
No log 5.6444 254 0.7225 0.7111 0.7225 0.8500
No log 5.6889 256 0.7842 0.7013 0.7842 0.8856
No log 5.7333 258 1.0404 0.6867 1.0404 1.0200
No log 5.7778 260 1.1630 0.6316 1.1630 1.0784
No log 5.8222 262 1.1507 0.5806 1.1507 1.0727
No log 5.8667 264 1.1190 0.5578 1.1190 1.0578
No log 5.9111 266 1.1112 0.5638 1.1112 1.0541
No log 5.9556 268 1.0439 0.6275 1.0439 1.0217
No log 6.0 270 0.9427 0.6383 0.9427 0.9709
No log 6.0444 272 0.9451 0.6475 0.9451 0.9722
No log 6.0889 274 0.9261 0.6429 0.9261 0.9624
No log 6.1333 276 0.9420 0.6569 0.9420 0.9706
No log 6.1778 278 0.9899 0.6525 0.9899 0.9949
No log 6.2222 280 1.0210 0.6483 1.0210 1.0105
No log 6.2667 282 0.9655 0.6383 0.9655 0.9826
No log 6.3111 284 0.9025 0.6571 0.9025 0.9500
No log 6.3556 286 0.8007 0.7194 0.8007 0.8948
No log 6.4 288 0.7703 0.7092 0.7703 0.8777
No log 6.4444 290 0.8747 0.6883 0.8747 0.9352
No log 6.4889 292 0.9783 0.6125 0.9783 0.9891
No log 6.5333 294 0.9473 0.6164 0.9473 0.9733
No log 6.5778 296 0.8496 0.6897 0.8496 0.9218
No log 6.6222 298 0.7770 0.7101 0.7770 0.8815
No log 6.6667 300 0.7920 0.6957 0.7920 0.8900
No log 6.7111 302 0.8651 0.6767 0.8651 0.9301
No log 6.7556 304 1.0292 0.6099 1.0292 1.0145
No log 6.8 306 1.1405 0.5578 1.1405 1.0679
No log 6.8444 308 1.1408 0.5867 1.1408 1.0681
No log 6.8889 310 1.0243 0.5714 1.0243 1.0121
No log 6.9333 312 0.9095 0.6933 0.9095 0.9537
No log 6.9778 314 0.8408 0.7320 0.8408 0.9169
No log 7.0222 316 0.7773 0.7368 0.7773 0.8816
No log 7.0667 318 0.7808 0.7368 0.7808 0.8836
No log 7.1111 320 0.8875 0.6708 0.8875 0.9421
No log 7.1556 322 0.9260 0.6289 0.9260 0.9623
No log 7.2 324 0.8588 0.7059 0.8588 0.9267
No log 7.2444 326 0.8459 0.7260 0.8459 0.9198
No log 7.2889 328 0.8059 0.7260 0.8059 0.8977
No log 7.3333 330 0.7671 0.7297 0.7671 0.8758
No log 7.3778 332 0.6828 0.7361 0.6828 0.8263
No log 7.4222 334 0.6656 0.7123 0.6656 0.8158
No log 7.4667 336 0.7241 0.7532 0.7241 0.8509
No log 7.5111 338 0.7413 0.7333 0.7413 0.8610
No log 7.5556 340 0.7043 0.7467 0.7043 0.8393
No log 7.6 342 0.7890 0.7237 0.7890 0.8883
No log 7.6444 344 0.8698 0.7190 0.8698 0.9326
No log 7.6889 346 0.9866 0.6752 0.9866 0.9933
No log 7.7333 348 1.0448 0.6752 1.0448 1.0221
No log 7.7778 350 1.0090 0.6797 1.0090 1.0045
No log 7.8222 352 0.8604 0.7027 0.8604 0.9276
No log 7.8667 354 0.7406 0.7153 0.7406 0.8606
No log 7.9111 356 0.7131 0.6917 0.7131 0.8444
No log 7.9556 358 0.7232 0.7059 0.7232 0.8504
No log 8.0 360 0.8268 0.6980 0.8268 0.9093
No log 8.0444 362 1.0296 0.6832 1.0296 1.0147
No log 8.0889 364 1.1153 0.6279 1.1153 1.0561
No log 8.1333 366 1.0060 0.6323 1.0060 1.0030
No log 8.1778 368 0.8334 0.6912 0.8334 0.9129
No log 8.2222 370 0.8341 0.7007 0.8341 0.9133
No log 8.2667 372 0.8298 0.6615 0.8298 0.9110
No log 8.3111 374 0.8100 0.6822 0.8100 0.9000
No log 8.3556 376 0.7757 0.6719 0.7757 0.8807
No log 8.4 378 0.7709 0.7068 0.7709 0.8780
No log 8.4444 380 0.8260 0.6763 0.8260 0.9088
No log 8.4889 382 0.8672 0.6475 0.8672 0.9312
No log 8.5333 384 0.8504 0.6571 0.8504 0.9222
No log 8.5778 386 0.8278 0.6857 0.8278 0.9098
No log 8.6222 388 0.7658 0.7007 0.7658 0.8751
No log 8.6667 390 0.7134 0.7465 0.7134 0.8446
No log 8.7111 392 0.7132 0.7465 0.7132 0.8445
No log 8.7556 394 0.7764 0.7133 0.7764 0.8811
No log 8.8 396 0.9500 0.6755 0.9500 0.9747
No log 8.8444 398 0.9296 0.6761 0.9296 0.9642
No log 8.8889 400 0.8074 0.7143 0.8074 0.8985
No log 8.9333 402 0.7891 0.7007 0.7891 0.8883
No log 8.9778 404 0.7659 0.7007 0.7659 0.8752
No log 9.0222 406 0.8090 0.7143 0.8090 0.8994
No log 9.0667 408 0.8610 0.6763 0.8610 0.9279
No log 9.1111 410 0.9038 0.6667 0.9038 0.9507
No log 9.1556 412 0.7972 0.7143 0.7972 0.8929
No log 9.2 414 0.7625 0.6912 0.7625 0.8732
No log 9.2444 416 0.7971 0.7143 0.7971 0.8928
No log 9.2889 418 0.9100 0.6667 0.9100 0.9539
No log 9.3333 420 1.0373 0.6056 1.0373 1.0185
No log 9.3778 422 1.0641 0.6093 1.0641 1.0316
No log 9.4222 424 0.9871 0.6345 0.9871 0.9935
No log 9.4667 426 0.8438 0.7042 0.8438 0.9186
No log 9.5111 428 0.8114 0.7133 0.8114 0.9008
No log 9.5556 430 0.8526 0.6897 0.8526 0.9234
No log 9.6 432 0.9044 0.6533 0.9044 0.9510
No log 9.6444 434 1.0584 0.6410 1.0584 1.0288
No log 9.6889 436 1.0944 0.6369 1.0944 1.0461
No log 9.7333 438 0.9952 0.6479 0.9952 0.9976
No log 9.7778 440 0.8977 0.6901 0.8977 0.9475
No log 9.8222 442 0.8780 0.6765 0.8780 0.9370
No log 9.8667 444 0.9215 0.6812 0.9215 0.9599
No log 9.9111 446 0.9981 0.6389 0.9981 0.9990
No log 9.9556 448 1.0200 0.6389 1.0200 1.0100
No log 10.0 450 0.9336 0.6763 0.9336 0.9663
No log 10.0444 452 0.8536 0.6667 0.8536 0.9239
No log 10.0889 454 0.7956 0.7015 0.7956 0.8920
No log 10.1333 456 0.8175 0.6812 0.8175 0.9042
No log 10.1778 458 0.9434 0.6712 0.9434 0.9713
No log 10.2222 460 1.1825 0.55 1.1825 1.0874
No log 10.2667 462 1.2015 0.5185 1.2015 1.0961
No log 10.3111 464 1.0699 0.6194 1.0699 1.0344
No log 10.3556 466 1.0798 0.5850 1.0798 1.0391
No log 10.4 468 1.0625 0.6525 1.0625 1.0308
No log 10.4444 470 0.9831 0.6466 0.9831 0.9915
No log 10.4889 472 0.9143 0.6718 0.9143 0.9562
No log 10.5333 474 0.8764 0.6718 0.8764 0.9362
No log 10.5778 476 0.9298 0.6466 0.9298 0.9642
No log 10.6222 478 1.0987 0.5890 1.0987 1.0482
No log 10.6667 480 1.1995 0.5290 1.1995 1.0952
No log 10.7111 482 1.1123 0.5419 1.1123 1.0547
No log 10.7556 484 1.0062 0.5755 1.0062 1.0031
No log 10.8 486 0.9532 0.6466 0.9532 0.9763
No log 10.8444 488 0.8982 0.6357 0.8982 0.9477
No log 10.8889 490 0.9303 0.6866 0.9303 0.9645
No log 10.9333 492 0.9783 0.6618 0.9783 0.9891
No log 10.9778 494 0.9428 0.6812 0.9428 0.9710
No log 11.0222 496 0.8859 0.6812 0.8859 0.9412
No log 11.0667 498 0.8932 0.6812 0.8932 0.9451
0.4073 11.1111 500 0.8486 0.6667 0.8486 0.9212
0.4073 11.1556 502 0.8541 0.6667 0.8541 0.9242
0.4073 11.2 504 0.8258 0.7111 0.8258 0.9087
0.4073 11.2444 506 0.8125 0.7111 0.8125 0.9014
0.4073 11.2889 508 0.8585 0.7050 0.8585 0.9266
0.4073 11.3333 510 0.9196 0.6573 0.9196 0.9590
0.4073 11.3778 512 0.9628 0.6577 0.9628 0.9812
0.4073 11.4222 514 0.9781 0.6839 0.9781 0.9890
0.4073 11.4667 516 0.9180 0.6846 0.9180 0.9581
0.4073 11.5111 518 0.7855 0.6944 0.7855 0.8863
0.4073 11.5556 520 0.7338 0.7133 0.7338 0.8566
0.4073 11.6 522 0.7880 0.6939 0.7880 0.8877
0.4073 11.6444 524 0.8488 0.6939 0.8488 0.9213
0.4073 11.6889 526 0.9607 0.6667 0.9607 0.9801
0.4073 11.7333 528 0.9340 0.6711 0.9340 0.9664
0.4073 11.7778 530 0.8450 0.6712 0.8450 0.9192
0.4073 11.8222 532 0.7709 0.6853 0.7709 0.8780
0.4073 11.8667 534 0.7931 0.6853 0.7931 0.8906
0.4073 11.9111 536 0.8990 0.6486 0.8990 0.9481
0.4073 11.9556 538 1.0464 0.6329 1.0464 1.0230
0.4073 12.0 540 1.1447 0.5848 1.1447 1.0699
0.4073 12.0444 542 1.1149 0.5854 1.1149 1.0559
0.4073 12.0889 544 1.0184 0.6581 1.0184 1.0092

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k6_task1_organization

Finetuned
(4023)
this model