ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k3_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5725
  • Qwk: 0.5607
  • Mse: 0.5725
  • Rmse: 0.7567

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1333 2 4.2058 -0.0177 4.2058 2.0508
No log 0.2667 4 3.2834 -0.0036 3.2834 1.8120
No log 0.4 6 1.3955 0.0760 1.3955 1.1813
No log 0.5333 8 1.1705 0.0967 1.1705 1.0819
No log 0.6667 10 1.0616 0.3541 1.0616 1.0304
No log 0.8 12 1.0069 0.2643 1.0069 1.0034
No log 0.9333 14 0.9937 0.2740 0.9937 0.9969
No log 1.0667 16 0.9431 0.2944 0.9431 0.9711
No log 1.2 18 0.9770 0.3935 0.9770 0.9884
No log 1.3333 20 1.1137 0.1498 1.1137 1.0553
No log 1.4667 22 1.0486 0.2969 1.0486 1.0240
No log 1.6 24 0.9977 0.3003 0.9977 0.9988
No log 1.7333 26 1.0085 0.2787 1.0085 1.0042
No log 1.8667 28 1.0653 0.1962 1.0653 1.0321
No log 2.0 30 1.0453 0.2175 1.0453 1.0224
No log 2.1333 32 0.9974 0.2175 0.9974 0.9987
No log 2.2667 34 0.8223 0.3883 0.8223 0.9068
No log 2.4 36 0.7608 0.4733 0.7608 0.8722
No log 2.5333 38 0.7471 0.5348 0.7471 0.8643
No log 2.6667 40 0.7775 0.4444 0.7775 0.8818
No log 2.8 42 0.9123 0.4462 0.9123 0.9551
No log 2.9333 44 0.9203 0.4094 0.9203 0.9593
No log 3.0667 46 0.8834 0.3208 0.8834 0.9399
No log 3.2 48 0.7497 0.5343 0.7497 0.8658
No log 3.3333 50 0.6601 0.5704 0.6601 0.8125
No log 3.4667 52 0.7183 0.5141 0.7183 0.8475
No log 3.6 54 0.6538 0.5225 0.6538 0.8086
No log 3.7333 56 0.6184 0.6073 0.6184 0.7864
No log 3.8667 58 0.6722 0.6043 0.6722 0.8199
No log 4.0 60 0.6707 0.5771 0.6707 0.8190
No log 4.1333 62 0.6482 0.6239 0.6482 0.8051
No log 4.2667 64 0.6322 0.5969 0.6322 0.7951
No log 4.4 66 0.6335 0.6073 0.6335 0.7959
No log 4.5333 68 0.6238 0.5690 0.6238 0.7898
No log 4.6667 70 0.6038 0.6498 0.6038 0.7771
No log 4.8 72 0.6148 0.6237 0.6148 0.7841
No log 4.9333 74 0.5941 0.6206 0.5941 0.7708
No log 5.0667 76 0.5846 0.6319 0.5846 0.7646
No log 5.2 78 0.5746 0.6484 0.5746 0.7580
No log 5.3333 80 0.5699 0.6484 0.5699 0.7549
No log 5.4667 82 0.6716 0.6766 0.6716 0.8195
No log 5.6 84 0.7061 0.6487 0.7061 0.8403
No log 5.7333 86 0.6205 0.6647 0.6205 0.7877
No log 5.8667 88 0.5941 0.6288 0.5941 0.7708
No log 6.0 90 0.6086 0.6619 0.6086 0.7801
No log 6.1333 92 0.6030 0.6439 0.6030 0.7765
No log 6.2667 94 0.6141 0.6664 0.6141 0.7836
No log 6.4 96 0.6500 0.6519 0.6500 0.8062
No log 6.5333 98 0.5991 0.6439 0.5991 0.7740
No log 6.6667 100 0.6770 0.6669 0.6770 0.8228
No log 6.8 102 0.6588 0.6548 0.6588 0.8117
No log 6.9333 104 0.5573 0.6545 0.5573 0.7465
No log 7.0667 106 0.7229 0.6128 0.7229 0.8502
No log 7.2 108 0.7203 0.6218 0.7203 0.8487
No log 7.3333 110 0.5605 0.6940 0.5605 0.7487
No log 7.4667 112 0.5271 0.6217 0.5271 0.7260
No log 7.6 114 0.5240 0.6916 0.5240 0.7238
No log 7.7333 116 0.5355 0.6424 0.5355 0.7317
No log 7.8667 118 0.5782 0.6500 0.5782 0.7604
No log 8.0 120 0.5890 0.6748 0.5890 0.7675
No log 8.1333 122 0.5422 0.6642 0.5422 0.7363
No log 8.2667 124 0.5571 0.6426 0.5571 0.7464
No log 8.4 126 0.5667 0.6606 0.5667 0.7528
No log 8.5333 128 0.5348 0.6642 0.5348 0.7313
No log 8.6667 130 0.6792 0.6304 0.6792 0.8241
No log 8.8 132 0.7280 0.6160 0.7280 0.8532
No log 8.9333 134 0.6795 0.4642 0.6795 0.8243
No log 9.0667 136 0.7695 0.5614 0.7695 0.8772
No log 9.2 138 0.6862 0.5829 0.6862 0.8284
No log 9.3333 140 0.5596 0.6465 0.5596 0.7481
No log 9.4667 142 0.5339 0.6614 0.5339 0.7307
No log 9.6 144 0.5344 0.6846 0.5344 0.7310
No log 9.7333 146 0.5646 0.6639 0.5646 0.7514
No log 9.8667 148 0.6631 0.6141 0.6631 0.8143
No log 10.0 150 0.6507 0.6141 0.6507 0.8067
No log 10.1333 152 0.5561 0.6371 0.5561 0.7457
No log 10.2667 154 0.5649 0.6473 0.5649 0.7516
No log 10.4 156 0.6344 0.6493 0.6344 0.7965
No log 10.5333 158 0.6055 0.6782 0.6055 0.7782
No log 10.6667 160 0.5454 0.6536 0.5454 0.7385
No log 10.8 162 0.5457 0.6822 0.5457 0.7387
No log 10.9333 164 0.6129 0.5593 0.6129 0.7829
No log 11.0667 166 0.6598 0.5527 0.6598 0.8123
No log 11.2 168 0.6002 0.5708 0.6002 0.7747
No log 11.3333 170 0.5377 0.6319 0.5377 0.7333
No log 11.4667 172 0.5588 0.6923 0.5588 0.7475
No log 11.6 174 0.5359 0.6923 0.5359 0.7320
No log 11.7333 176 0.5339 0.7106 0.5339 0.7307
No log 11.8667 178 0.5738 0.6395 0.5738 0.7575
No log 12.0 180 0.5694 0.6218 0.5694 0.7546
No log 12.1333 182 0.5524 0.7059 0.5524 0.7432
No log 12.2667 184 0.5483 0.6400 0.5483 0.7405
No log 12.4 186 0.5261 0.7026 0.5261 0.7254
No log 12.5333 188 0.5329 0.6328 0.5329 0.7300
No log 12.6667 190 0.5238 0.6724 0.5238 0.7237
No log 12.8 192 0.5210 0.6909 0.5210 0.7218
No log 12.9333 194 0.5209 0.6766 0.5209 0.7217
No log 13.0667 196 0.5590 0.6275 0.5590 0.7476
No log 13.2 198 0.6286 0.6597 0.6286 0.7929
No log 13.3333 200 0.5886 0.6275 0.5886 0.7672
No log 13.4667 202 0.5413 0.6032 0.5413 0.7357
No log 13.6 204 0.5629 0.6990 0.5629 0.7503
No log 13.7333 206 0.5507 0.6215 0.5507 0.7421
No log 13.8667 208 0.6228 0.6698 0.6228 0.7892
No log 14.0 210 0.6987 0.6712 0.6987 0.8359
No log 14.1333 212 0.6210 0.6035 0.6210 0.7881
No log 14.2667 214 0.5849 0.6157 0.5849 0.7648
No log 14.4 216 0.5940 0.6147 0.5940 0.7707
No log 14.5333 218 0.6007 0.6147 0.6007 0.7750
No log 14.6667 220 0.5792 0.6473 0.5792 0.7611
No log 14.8 222 0.5667 0.6536 0.5667 0.7528
No log 14.9333 224 0.5852 0.6087 0.5852 0.7650
No log 15.0667 226 0.5978 0.5961 0.5978 0.7731
No log 15.2 228 0.5716 0.6018 0.5716 0.7561
No log 15.3333 230 0.5898 0.6167 0.5898 0.7680
No log 15.4667 232 0.6125 0.5833 0.6125 0.7826
No log 15.6 234 0.6425 0.5938 0.6425 0.8016
No log 15.7333 236 0.5954 0.5921 0.5954 0.7716
No log 15.8667 238 0.5957 0.5933 0.5957 0.7718
No log 16.0 240 0.6549 0.6259 0.6549 0.8092
No log 16.1333 242 0.5984 0.6334 0.5984 0.7736
No log 16.2667 244 0.5636 0.6078 0.5636 0.7507
No log 16.4 246 0.5829 0.5948 0.5829 0.7635
No log 16.5333 248 0.6378 0.5799 0.6378 0.7986
No log 16.6667 250 0.6001 0.5510 0.6001 0.7747
No log 16.8 252 0.5855 0.6320 0.5855 0.7652
No log 16.9333 254 0.5883 0.6667 0.5883 0.7670
No log 17.0667 256 0.5653 0.6672 0.5653 0.7519
No log 17.2 258 0.5983 0.6404 0.5983 0.7735
No log 17.3333 260 0.6478 0.5973 0.6478 0.8048
No log 17.4667 262 0.6520 0.5777 0.6520 0.8075
No log 17.6 264 0.5950 0.6025 0.5950 0.7714
No log 17.7333 266 0.5918 0.5948 0.5918 0.7693
No log 17.8667 268 0.5916 0.5645 0.5916 0.7692
No log 18.0 270 0.6474 0.5833 0.6474 0.8046
No log 18.1333 272 0.6947 0.5591 0.6947 0.8335
No log 18.2667 274 0.6459 0.5998 0.6459 0.8037
No log 18.4 276 0.5546 0.5959 0.5546 0.7447
No log 18.5333 278 0.5282 0.6748 0.5282 0.7268
No log 18.6667 280 0.5226 0.6642 0.5226 0.7229
No log 18.8 282 0.5690 0.6830 0.5690 0.7544
No log 18.9333 284 0.6135 0.6690 0.6135 0.7833
No log 19.0667 286 0.5948 0.6573 0.5948 0.7712
No log 19.2 288 0.5611 0.6237 0.5611 0.7491
No log 19.3333 290 0.5533 0.6745 0.5533 0.7439
No log 19.4667 292 0.5763 0.6102 0.5763 0.7591
No log 19.6 294 0.5925 0.5975 0.5925 0.7697
No log 19.7333 296 0.5602 0.6441 0.5602 0.7485
No log 19.8667 298 0.5425 0.6301 0.5425 0.7365
No log 20.0 300 0.5591 0.6147 0.5591 0.7478
No log 20.1333 302 0.5518 0.6647 0.5518 0.7428
No log 20.2667 304 0.5290 0.6347 0.5290 0.7273
No log 20.4 306 0.5504 0.6328 0.5504 0.7419
No log 20.5333 308 0.6119 0.6459 0.6119 0.7822
No log 20.6667 310 0.6111 0.6287 0.6111 0.7817
No log 20.8 312 0.5906 0.6109 0.5906 0.7685
No log 20.9333 314 0.5453 0.6641 0.5453 0.7385
No log 21.0667 316 0.5547 0.6437 0.5547 0.7448
No log 21.2 318 0.5567 0.6575 0.5567 0.7461
No log 21.3333 320 0.5876 0.5833 0.5876 0.7666
No log 21.4667 322 0.6894 0.6209 0.6894 0.8303
No log 21.6 324 0.7613 0.6078 0.7613 0.8725
No log 21.7333 326 0.7426 0.6377 0.7426 0.8617
No log 21.8667 328 0.6634 0.6141 0.6634 0.8145
No log 22.0 330 0.5796 0.6073 0.5796 0.7613
No log 22.1333 332 0.5845 0.6133 0.5845 0.7645
No log 22.2667 334 0.5865 0.6133 0.5865 0.7659
No log 22.4 336 0.5878 0.6102 0.5878 0.7667
No log 22.5333 338 0.6949 0.5666 0.6949 0.8336
No log 22.6667 340 0.8333 0.5966 0.8333 0.9129
No log 22.8 342 0.8298 0.5966 0.8298 0.9109
No log 22.9333 344 0.7184 0.6035 0.7184 0.8476
No log 23.0667 346 0.6143 0.6063 0.6143 0.7838
No log 23.2 348 0.5761 0.6430 0.5761 0.7590
No log 23.3333 350 0.5737 0.6430 0.5737 0.7574
No log 23.4667 352 0.5921 0.6446 0.5921 0.7695
No log 23.6 354 0.6558 0.5487 0.6558 0.8098
No log 23.7333 356 0.6905 0.5666 0.6905 0.8310
No log 23.8667 358 0.6410 0.5697 0.6410 0.8007
No log 24.0 360 0.5942 0.6035 0.5942 0.7708
No log 24.1333 362 0.5809 0.6177 0.5809 0.7622
No log 24.2667 364 0.5549 0.6593 0.5549 0.7449
No log 24.4 366 0.5544 0.6593 0.5544 0.7446
No log 24.5333 368 0.5661 0.6473 0.5661 0.7524
No log 24.6667 370 0.6078 0.5602 0.6078 0.7796
No log 24.8 372 0.6206 0.5487 0.6206 0.7878
No log 24.9333 374 0.6383 0.5487 0.6383 0.7989
No log 25.0667 376 0.6240 0.5487 0.6240 0.7899
No log 25.2 378 0.5837 0.5833 0.5837 0.7640
No log 25.3333 380 0.5767 0.6167 0.5767 0.7594
No log 25.4667 382 0.5548 0.6282 0.5548 0.7449
No log 25.6 384 0.5408 0.6396 0.5408 0.7354
No log 25.7333 386 0.5542 0.6282 0.5542 0.7444
No log 25.8667 388 0.5845 0.5988 0.5845 0.7645
No log 26.0 390 0.6206 0.5988 0.6206 0.7878
No log 26.1333 392 0.6231 0.5877 0.6231 0.7894
No log 26.2667 394 0.6055 0.5988 0.6055 0.7781
No log 26.4 396 0.5600 0.6063 0.5600 0.7483
No log 26.5333 398 0.5291 0.6712 0.5291 0.7274
No log 26.6667 400 0.5281 0.6509 0.5281 0.7267
No log 26.8 402 0.5235 0.6572 0.5235 0.7235
No log 26.9333 404 0.5392 0.6396 0.5392 0.7343
No log 27.0667 406 0.5803 0.6319 0.5803 0.7618
No log 27.2 408 0.6282 0.5877 0.6282 0.7926
No log 27.3333 410 0.6341 0.5602 0.6341 0.7963
No log 27.4667 412 0.6180 0.5602 0.6180 0.7861
No log 27.6 414 0.5901 0.5948 0.5901 0.7682
No log 27.7333 416 0.5687 0.6241 0.5687 0.7541
No log 27.8667 418 0.5599 0.6113 0.5599 0.7483
No log 28.0 420 0.5525 0.6227 0.5525 0.7433
No log 28.1333 422 0.5402 0.6732 0.5402 0.7350
No log 28.2667 424 0.5756 0.6493 0.5756 0.7587
No log 28.4 426 0.6073 0.6071 0.6073 0.7793
No log 28.5333 428 0.6348 0.6071 0.6348 0.7968
No log 28.6667 430 0.6452 0.5718 0.6452 0.8032
No log 28.8 432 0.6369 0.5718 0.6369 0.7981
No log 28.9333 434 0.6063 0.6053 0.6063 0.7786
No log 29.0667 436 0.5912 0.6167 0.5912 0.7689
No log 29.2 438 0.5724 0.6464 0.5724 0.7566
No log 29.3333 440 0.5666 0.6464 0.5666 0.7528
No log 29.4667 442 0.6067 0.6247 0.6067 0.7789
No log 29.6 444 0.6109 0.6061 0.6109 0.7816
No log 29.7333 446 0.5830 0.5799 0.5830 0.7636
No log 29.8667 448 0.5461 0.6113 0.5461 0.7390
No log 30.0 450 0.5525 0.6466 0.5525 0.7433
No log 30.1333 452 0.5546 0.6657 0.5546 0.7447
No log 30.2667 454 0.5579 0.6025 0.5579 0.7469
No log 30.4 456 0.5750 0.5912 0.5750 0.7583
No log 30.5333 458 0.5680 0.6352 0.5680 0.7537
No log 30.6667 460 0.5530 0.6748 0.5530 0.7436
No log 30.8 462 0.5546 0.6748 0.5546 0.7447
No log 30.9333 464 0.5425 0.6259 0.5425 0.7366
No log 31.0667 466 0.5467 0.6259 0.5467 0.7394
No log 31.2 468 0.5646 0.6575 0.5646 0.7514
No log 31.3333 470 0.5954 0.6127 0.5954 0.7716
No log 31.4667 472 0.5932 0.6240 0.5932 0.7702
No log 31.6 474 0.5794 0.6352 0.5794 0.7612
No log 31.7333 476 0.5599 0.6575 0.5599 0.7483
No log 31.8667 478 0.5447 0.6259 0.5447 0.7381
No log 32.0 480 0.5473 0.6259 0.5473 0.7398
No log 32.1333 482 0.5836 0.6529 0.5836 0.7639
No log 32.2667 484 0.6213 0.6278 0.6213 0.7882
No log 32.4 486 0.6192 0.6099 0.6192 0.7869
No log 32.5333 488 0.5764 0.6464 0.5764 0.7592
No log 32.6667 490 0.5489 0.6219 0.5489 0.7409
No log 32.8 492 0.5572 0.6286 0.5572 0.7465
No log 32.9333 494 0.5649 0.6241 0.5649 0.7516
No log 33.0667 496 0.5987 0.5948 0.5987 0.7737
No log 33.2 498 0.6847 0.5877 0.6847 0.8275
0.2491 33.3333 500 0.7345 0.5943 0.7345 0.8570
0.2491 33.4667 502 0.7043 0.5766 0.7043 0.8392
0.2491 33.6 504 0.6132 0.5912 0.6132 0.7830
0.2491 33.7333 506 0.5464 0.6479 0.5464 0.7392
0.2491 33.8667 508 0.5488 0.6556 0.5488 0.7408
0.2491 34.0 510 0.5586 0.6447 0.5586 0.7474
0.2491 34.1333 512 0.5467 0.6374 0.5467 0.7394
0.2491 34.2667 514 0.5415 0.6415 0.5415 0.7359
0.2491 34.4 516 0.5832 0.6035 0.5832 0.7637
0.2491 34.5333 518 0.6146 0.6310 0.6146 0.7839
0.2491 34.6667 520 0.6007 0.6147 0.6007 0.7751
0.2491 34.8 522 0.5829 0.5959 0.5829 0.7635
0.2491 34.9333 524 0.5739 0.5959 0.5739 0.7576
0.2491 35.0667 526 0.5632 0.6065 0.5632 0.7505
0.2491 35.2 528 0.5556 0.6035 0.5556 0.7454
0.2491 35.3333 530 0.5519 0.5969 0.5519 0.7429
0.2491 35.4667 532 0.5710 0.6352 0.5710 0.7556
0.2491 35.6 534 0.5738 0.6352 0.5738 0.7575
0.2491 35.7333 536 0.5707 0.6249 0.5707 0.7555
0.2491 35.8667 538 0.5637 0.6249 0.5637 0.7508
0.2491 36.0 540 0.5625 0.6249 0.5625 0.7500
0.2491 36.1333 542 0.5631 0.5986 0.5631 0.7504
0.2491 36.2667 544 0.5641 0.5986 0.5641 0.7511
0.2491 36.4 546 0.5790 0.6500 0.5790 0.7609
0.2491 36.5333 548 0.6171 0.5799 0.6171 0.7856
0.2491 36.6667 550 0.6388 0.5686 0.6388 0.7993
0.2491 36.8 552 0.6172 0.5877 0.6172 0.7856
0.2491 36.9333 554 0.6027 0.5877 0.6027 0.7763
0.2491 37.0667 556 0.5725 0.5607 0.5725 0.7567

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k3_task5_organization

Finetuned
(4019)
this model