ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k4_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0011
  • Qwk: 0.6074
  • Mse: 1.0011
  • Rmse: 1.0006

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0667 2 6.9764 0.0171 6.9764 2.6413
No log 0.1333 4 4.8082 0.0682 4.8082 2.1928
No log 0.2 6 2.9514 0.0952 2.9514 1.7180
No log 0.2667 8 2.1080 0.1642 2.1080 1.4519
No log 0.3333 10 1.9846 0.1368 1.9846 1.4088
No log 0.4 12 1.9140 0.1770 1.9140 1.3835
No log 0.4667 14 1.5658 0.1538 1.5658 1.2513
No log 0.5333 16 1.4972 0.1333 1.4972 1.2236
No log 0.6 18 1.5269 0.1165 1.5269 1.2357
No log 0.6667 20 1.4226 0.1698 1.4226 1.1927
No log 0.7333 22 1.4216 0.2222 1.4216 1.1923
No log 0.8 24 1.3511 0.2364 1.3511 1.1623
No log 0.8667 26 1.2903 0.2909 1.2903 1.1359
No log 0.9333 28 1.3657 0.3243 1.3657 1.1686
No log 1.0 30 1.5292 0.2909 1.5292 1.2366
No log 1.0667 32 1.4034 0.3009 1.4034 1.1846
No log 1.1333 34 1.3252 0.3363 1.3252 1.1512
No log 1.2 36 1.2361 0.4724 1.2361 1.1118
No log 1.2667 38 1.2578 0.4154 1.2578 1.1215
No log 1.3333 40 1.4150 0.4889 1.4150 1.1895
No log 1.4 42 1.2796 0.4889 1.2796 1.1312
No log 1.4667 44 1.3044 0.3902 1.3044 1.1421
No log 1.5333 46 1.3404 0.4 1.3404 1.1577
No log 1.6 48 1.1909 0.5873 1.1909 1.0913
No log 1.6667 50 1.1548 0.5691 1.1548 1.0746
No log 1.7333 52 1.3146 0.4715 1.3146 1.1465
No log 1.8 54 1.4427 0.3817 1.4427 1.2011
No log 1.8667 56 1.4663 0.3788 1.4663 1.2109
No log 1.9333 58 1.3162 0.5333 1.3162 1.1473
No log 2.0 60 1.1685 0.5362 1.1685 1.0810
No log 2.0667 62 1.1713 0.5109 1.1713 1.0823
No log 2.1333 64 1.0714 0.5441 1.0714 1.0351
No log 2.2 66 0.9446 0.6308 0.9446 0.9719
No log 2.2667 68 0.8931 0.6364 0.8931 0.9451
No log 2.3333 70 0.8875 0.6212 0.8875 0.9421
No log 2.4 72 0.9319 0.5692 0.9319 0.9654
No log 2.4667 74 1.1671 0.5839 1.1671 1.0803
No log 2.5333 76 1.1438 0.5926 1.1438 1.0695
No log 2.6 78 0.9313 0.5882 0.9313 0.9650
No log 2.6667 80 0.8989 0.6479 0.8989 0.9481
No log 2.7333 82 1.0399 0.6364 1.0399 1.0198
No log 2.8 84 1.1140 0.5974 1.1140 1.0555
No log 2.8667 86 1.0038 0.5793 1.0038 1.0019
No log 2.9333 88 0.8592 0.6573 0.8592 0.9269
No log 3.0 90 0.8203 0.6993 0.8203 0.9057
No log 3.0667 92 0.8649 0.6853 0.8649 0.9300
No log 3.1333 94 0.8537 0.6957 0.8537 0.9239
No log 3.2 96 1.0046 0.5797 1.0046 1.0023
No log 3.2667 98 1.1595 0.5072 1.1595 1.0768
No log 3.3333 100 1.0644 0.5147 1.0644 1.0317
No log 3.4 102 0.9414 0.5630 0.9414 0.9703
No log 3.4667 104 0.9132 0.6212 0.9131 0.9556
No log 3.5333 106 0.9298 0.6519 0.9298 0.9642
No log 3.6 108 0.9704 0.6567 0.9704 0.9851
No log 3.6667 110 0.9746 0.6567 0.9746 0.9872
No log 3.7333 112 0.9745 0.5735 0.9745 0.9872
No log 3.8 114 0.9215 0.6131 0.9215 0.9599
No log 3.8667 116 0.9143 0.6569 0.9143 0.9562
No log 3.9333 118 0.9797 0.5775 0.9797 0.9898
No log 4.0 120 1.1843 0.5732 1.1843 1.0883
No log 4.0667 122 1.2475 0.5614 1.2475 1.1169
No log 4.1333 124 1.1198 0.5769 1.1198 1.0582
No log 4.2 126 0.9862 0.6309 0.9862 0.9931
No log 4.2667 128 0.8883 0.6761 0.8883 0.9425
No log 4.3333 130 0.9083 0.6475 0.9083 0.9530
No log 4.4 132 0.9153 0.6667 0.9153 0.9567
No log 4.4667 134 0.9408 0.6525 0.9408 0.9700
No log 4.5333 136 1.0960 0.5931 1.0960 1.0469
No log 4.6 138 1.3627 0.4583 1.3627 1.1673
No log 4.6667 140 1.3957 0.4429 1.3957 1.1814
No log 4.7333 142 1.3212 0.4923 1.3212 1.1494
No log 4.8 144 1.2972 0.5 1.2972 1.1389
No log 4.8667 146 1.2585 0.4918 1.2585 1.1219
No log 4.9333 148 1.1776 0.4407 1.1776 1.0852
No log 5.0 150 1.0476 0.5289 1.0476 1.0235
No log 5.0667 152 0.9520 0.6364 0.9520 0.9757
No log 5.1333 154 0.9052 0.6338 0.9052 0.9514
No log 5.2 156 0.9445 0.6883 0.9445 0.9718
No log 5.2667 158 0.9969 0.6541 0.9969 0.9985
No log 5.3333 160 0.8501 0.7044 0.8501 0.9220
No log 5.4 162 0.7043 0.7730 0.7043 0.8392
No log 5.4667 164 0.6873 0.7550 0.6873 0.8291
No log 5.5333 166 0.7181 0.7347 0.7181 0.8474
No log 5.6 168 0.7281 0.7465 0.7281 0.8533
No log 5.6667 170 0.7835 0.7320 0.7835 0.8852
No log 5.7333 172 1.0219 0.6588 1.0219 1.0109
No log 5.8 174 1.2925 0.5889 1.2925 1.1369
No log 5.8667 176 1.2800 0.5263 1.2800 1.1314
No log 5.9333 178 1.1537 0.5106 1.1537 1.0741
No log 6.0 180 0.8899 0.6939 0.8899 0.9434
No log 6.0667 182 0.8726 0.6986 0.8726 0.9341
No log 6.1333 184 0.9254 0.6892 0.9254 0.9620
No log 6.2 186 0.9707 0.6486 0.9707 0.9853
No log 6.2667 188 0.9825 0.6667 0.9825 0.9912
No log 6.3333 190 1.0853 0.5571 1.0853 1.0418
No log 6.4 192 1.2074 0.5714 1.2074 1.0988
No log 6.4667 194 1.2344 0.5032 1.2344 1.1110
No log 6.5333 196 1.0983 0.5634 1.0983 1.0480
No log 6.6 198 0.8863 0.6615 0.8863 0.9414
No log 6.6667 200 0.8662 0.6260 0.8662 0.9307
No log 6.7333 202 0.8438 0.6515 0.8438 0.9186
No log 6.8 204 0.8758 0.6569 0.8758 0.9359
No log 6.8667 206 1.0794 0.5986 1.0794 1.0389
No log 6.9333 208 1.2339 0.5395 1.2339 1.1108
No log 7.0 210 1.1472 0.5616 1.1472 1.0711
No log 7.0667 212 1.0086 0.6316 1.0086 1.0043
No log 7.1333 214 0.9640 0.5873 0.9640 0.9818
No log 7.2 216 0.9627 0.56 0.9627 0.9812
No log 7.2667 218 0.9882 0.6094 0.9882 0.9941
No log 7.3333 220 1.0692 0.6212 1.0692 1.0340
No log 7.4 222 1.2080 0.5235 1.2080 1.0991
No log 7.4667 224 1.2231 0.5714 1.2231 1.1059
No log 7.5333 226 1.0384 0.5811 1.0384 1.0190
No log 7.6 228 0.8794 0.7153 0.8794 0.9377
No log 7.6667 230 0.8141 0.7391 0.8141 0.9023
No log 7.7333 232 0.7751 0.7376 0.7751 0.8804
No log 7.8 234 0.7910 0.7246 0.7910 0.8894
No log 7.8667 236 0.8405 0.7083 0.8405 0.9168
No log 7.9333 238 0.8750 0.6993 0.8750 0.9354
No log 8.0 240 0.9187 0.6345 0.9187 0.9585
No log 8.0667 242 0.9314 0.6301 0.9314 0.9651
No log 8.1333 244 0.8596 0.6809 0.8596 0.9272
No log 8.2 246 0.7642 0.7111 0.7642 0.8742
No log 8.2667 248 0.7321 0.6963 0.7321 0.8556
No log 8.3333 250 0.7325 0.6866 0.7325 0.8559
No log 8.4 252 0.8065 0.7007 0.8065 0.8980
No log 8.4667 254 0.9533 0.6479 0.9533 0.9764
No log 8.5333 256 1.0623 0.5931 1.0623 1.0307
No log 8.6 258 1.0365 0.6434 1.0365 1.0181
No log 8.6667 260 1.0304 0.6056 1.0304 1.0151
No log 8.7333 262 1.0869 0.5429 1.0869 1.0425
No log 8.8 264 1.1559 0.5753 1.1559 1.0751
No log 8.8667 266 1.1009 0.5985 1.1009 1.0493
No log 8.9333 268 1.0305 0.6119 1.0305 1.0151
No log 9.0 270 0.9879 0.6370 0.9879 0.9939
No log 9.0667 272 0.9734 0.6471 0.9734 0.9866
No log 9.1333 274 0.9319 0.6667 0.9319 0.9653
No log 9.2 276 0.9512 0.6277 0.9512 0.9753
No log 9.2667 278 1.0410 0.6623 1.0410 1.0203
No log 9.3333 280 1.1359 0.6173 1.1359 1.0658
No log 9.4 282 1.0596 0.6164 1.0596 1.0294
No log 9.4667 284 0.9570 0.6187 0.9570 0.9782
No log 9.5333 286 0.9058 0.6615 0.9058 0.9518
No log 9.6 288 0.9349 0.6515 0.9349 0.9669
No log 9.6667 290 1.0271 0.5926 1.0271 1.0135
No log 9.7333 292 1.1637 0.5278 1.1637 1.0788
No log 9.8 294 1.1803 0.5294 1.1803 1.0864
No log 9.8667 296 1.0889 0.5909 1.0889 1.0435
No log 9.9333 298 0.9854 0.6512 0.9854 0.9927
No log 10.0 300 0.9463 0.6406 0.9463 0.9728
No log 10.0667 302 0.9374 0.6512 0.9374 0.9682
No log 10.1333 304 0.9198 0.6308 0.9198 0.9590
No log 10.2 306 0.9174 0.6389 0.9174 0.9578
No log 10.2667 308 0.9304 0.6490 0.9304 0.9646
No log 10.3333 310 0.8553 0.6806 0.8553 0.9248
No log 10.4 312 0.8147 0.7259 0.8147 0.9026
No log 10.4667 314 0.8363 0.7353 0.8363 0.9145
No log 10.5333 316 0.8862 0.6389 0.8862 0.9414
No log 10.6 318 0.9660 0.6438 0.9660 0.9829
No log 10.6667 320 0.9600 0.6622 0.9600 0.9798
No log 10.7333 322 0.8847 0.6232 0.8847 0.9406
No log 10.8 324 0.8607 0.6718 0.8607 0.9278
No log 10.8667 326 0.8710 0.6716 0.8710 0.9333
No log 10.9333 328 0.9155 0.6512 0.9155 0.9568
No log 11.0 330 0.9735 0.6269 0.9735 0.9867
No log 11.0667 332 1.0916 0.5874 1.0916 1.0448
No log 11.1333 334 1.2164 0.5205 1.2164 1.1029
No log 11.2 336 1.1818 0.5333 1.1818 1.0871
No log 11.2667 338 1.0811 0.6615 1.0811 1.0398
No log 11.3333 340 0.9789 0.5984 0.9789 0.9894
No log 11.4 342 0.9074 0.5984 0.9074 0.9526
No log 11.4667 344 0.8539 0.6716 0.8539 0.9241
No log 11.5333 346 0.8375 0.6718 0.8375 0.9151
No log 11.6 348 0.8588 0.6767 0.8588 0.9267
No log 11.6667 350 0.9032 0.6861 0.9032 0.9504
No log 11.7333 352 0.8940 0.6765 0.8940 0.9455
No log 11.8 354 0.9132 0.6573 0.9132 0.9556
No log 11.8667 356 0.8525 0.7007 0.8525 0.9233
No log 11.9333 358 0.8249 0.7407 0.8249 0.9082
No log 12.0 360 0.7909 0.7164 0.7909 0.8893
No log 12.0667 362 0.7706 0.7059 0.7706 0.8778
No log 12.1333 364 0.7525 0.7153 0.7525 0.8675
No log 12.2 366 0.7903 0.7211 0.7903 0.8890
No log 12.2667 368 0.8736 0.6667 0.8736 0.9347
No log 12.3333 370 0.8924 0.6667 0.8924 0.9447
No log 12.4 372 0.8375 0.6567 0.8375 0.9151
No log 12.4667 374 0.8197 0.7015 0.8197 0.9054
No log 12.5333 376 0.8507 0.7068 0.8507 0.9223
No log 12.6 378 0.9027 0.6515 0.9027 0.9501
No log 12.6667 380 0.9673 0.6515 0.9673 0.9835
No log 12.7333 382 0.9843 0.6412 0.9843 0.9921
No log 12.8 384 0.9523 0.6202 0.9523 0.9758
No log 12.8667 386 0.8977 0.6462 0.8977 0.9475
No log 12.9333 388 0.8584 0.6667 0.8584 0.9265
No log 13.0 390 0.8525 0.6429 0.8525 0.9233
No log 13.0667 392 0.8266 0.6849 0.8266 0.9092
No log 13.1333 394 0.8232 0.6759 0.8232 0.9073
No log 13.2 396 0.8122 0.6620 0.8122 0.9012
No log 13.2667 398 0.8215 0.6809 0.8215 0.9063
No log 13.3333 400 0.8046 0.6917 0.8046 0.8970
No log 13.4 402 0.8135 0.6769 0.8135 0.9020
No log 13.4667 404 0.8524 0.6565 0.8524 0.9233
No log 13.5333 406 0.8953 0.6269 0.8953 0.9462
No log 13.6 408 0.9143 0.6277 0.9143 0.9562
No log 13.6667 410 0.8952 0.6716 0.8952 0.9461
No log 13.7333 412 0.9167 0.6667 0.9167 0.9575
No log 13.8 414 0.9323 0.6618 0.9323 0.9656
No log 13.8667 416 0.9338 0.6618 0.9338 0.9663
No log 13.9333 418 0.9310 0.6667 0.9310 0.9649
No log 14.0 420 0.9447 0.6767 0.9447 0.9720
No log 14.0667 422 0.9393 0.6615 0.9393 0.9692
No log 14.1333 424 0.9338 0.6615 0.9338 0.9663
No log 14.2 426 0.9310 0.6615 0.9310 0.9649
No log 14.2667 428 0.9273 0.6615 0.9273 0.9630
No log 14.3333 430 0.9021 0.6565 0.9021 0.9498
No log 14.4 432 0.8733 0.6715 0.8733 0.9345
No log 14.4667 434 0.8550 0.6809 0.8550 0.9247
No log 14.5333 436 0.8580 0.7042 0.8580 0.9263
No log 14.6 438 0.8947 0.6575 0.8947 0.9459
No log 14.6667 440 0.8974 0.6622 0.8974 0.9473
No log 14.7333 442 0.8713 0.6944 0.8713 0.9334
No log 14.8 444 0.8501 0.7246 0.8501 0.9220
No log 14.8667 446 0.8603 0.7059 0.8603 0.9275
No log 14.9333 448 0.8651 0.6718 0.8651 0.9301
No log 15.0 450 0.8960 0.6512 0.8960 0.9466
No log 15.0667 452 0.9191 0.6418 0.9191 0.9587
No log 15.1333 454 0.9090 0.6277 0.9090 0.9534
No log 15.2 456 0.8938 0.6383 0.8938 0.9454
No log 15.2667 458 0.8702 0.6479 0.8702 0.9328
No log 15.3333 460 0.8913 0.6528 0.8913 0.9441
No log 15.4 462 0.9148 0.6573 0.9148 0.9564
No log 15.4667 464 0.8861 0.6667 0.8861 0.9413
No log 15.5333 466 0.8409 0.6667 0.8409 0.9170
No log 15.6 468 0.8334 0.6923 0.8334 0.9129
No log 15.6667 470 0.8137 0.7023 0.8137 0.9020
No log 15.7333 472 0.7878 0.7121 0.7878 0.8876
No log 15.8 474 0.7980 0.7259 0.7980 0.8933
No log 15.8667 476 0.8337 0.6763 0.8337 0.9131
No log 15.9333 478 0.8482 0.6618 0.8482 0.9210
No log 16.0 480 0.8863 0.6519 0.8863 0.9415
No log 16.0667 482 0.8811 0.6618 0.8811 0.9387
No log 16.1333 484 0.8389 0.6615 0.8389 0.9159
No log 16.2 486 0.8194 0.7121 0.8194 0.9052
No log 16.2667 488 0.8232 0.6963 0.8232 0.9073
No log 16.3333 490 0.8243 0.6716 0.8243 0.9079
No log 16.4 492 0.8067 0.6957 0.8067 0.8982
No log 16.4667 494 0.7987 0.7310 0.7987 0.8937
No log 16.5333 496 0.8228 0.6849 0.8228 0.9071
No log 16.6 498 0.9020 0.6788 0.9020 0.9498
0.3718 16.6667 500 1.0266 0.6341 1.0266 1.0132
0.3718 16.7333 502 0.9694 0.6456 0.9694 0.9846
0.3718 16.8 504 0.8934 0.6573 0.8934 0.9452
0.3718 16.8667 506 0.9105 0.6241 0.9105 0.9542
0.3718 16.9333 508 0.9031 0.6412 0.9031 0.9503
0.3718 17.0 510 0.9229 0.6212 0.9229 0.9607
0.3718 17.0667 512 0.9255 0.6212 0.9255 0.9621
0.3718 17.1333 514 0.8999 0.6316 0.8999 0.9486
0.3718 17.2 516 0.8433 0.6718 0.8433 0.9183
0.3718 17.2667 518 0.8042 0.7164 0.8042 0.8968
0.3718 17.3333 520 0.8017 0.7164 0.8017 0.8954
0.3718 17.4 522 0.8428 0.6286 0.8428 0.9180
0.3718 17.4667 524 0.8718 0.6286 0.8718 0.9337
0.3718 17.5333 526 0.8672 0.6475 0.8672 0.9313
0.3718 17.6 528 0.8980 0.6522 0.8980 0.9476
0.3718 17.6667 530 0.9030 0.6269 0.9030 0.9502
0.3718 17.7333 532 0.9463 0.6074 0.9463 0.9728
0.3718 17.8 534 1.0030 0.6074 1.0030 1.0015
0.3718 17.8667 536 1.0320 0.6074 1.0320 1.0159
0.3718 17.9333 538 1.0011 0.6074 1.0011 1.0006

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k4_task1_organization

Finetuned
(4023)
this model