ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k14_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9169
  • Qwk: -0.2056
  • Mse: 0.9169
  • Rmse: 0.9576

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0571 2 3.8045 -0.0154 3.8045 1.9505
No log 0.1143 4 1.7555 0.0943 1.7555 1.3250
No log 0.1714 6 1.7680 0.0674 1.7680 1.3297
No log 0.2286 8 1.3604 -0.0145 1.3604 1.1664
No log 0.2857 10 1.3297 -0.0619 1.3297 1.1531
No log 0.3429 12 0.8868 -0.1253 0.8868 0.9417
No log 0.4 14 0.6477 0.1512 0.6477 0.8048
No log 0.4571 16 0.7049 0.0628 0.7049 0.8396
No log 0.5143 18 1.1486 0.0651 1.1486 1.0717
No log 0.5714 20 1.2589 -0.0736 1.2589 1.1220
No log 0.6286 22 1.2197 0.0032 1.2197 1.1044
No log 0.6857 24 1.2817 -0.0247 1.2817 1.1321
No log 0.7429 26 1.2978 0.0 1.2978 1.1392
No log 0.8 28 1.2029 0.0 1.2029 1.0968
No log 0.8571 30 1.0049 0.0 1.0049 1.0024
No log 0.9143 32 0.8542 -0.0583 0.8542 0.9242
No log 0.9714 34 0.8049 0.1318 0.8049 0.8971
No log 1.0286 36 0.8066 0.0628 0.8066 0.8981
No log 1.0857 38 0.9194 0.0100 0.9194 0.9589
No log 1.1429 40 0.8613 0.1456 0.8613 0.9280
No log 1.2 42 0.7545 0.1021 0.7545 0.8686
No log 1.2571 44 0.7784 0.0374 0.7784 0.8823
No log 1.3143 46 0.8127 0.0129 0.8127 0.9015
No log 1.3714 48 0.9344 -0.0207 0.9344 0.9666
No log 1.4286 50 1.4709 0.0 1.4709 1.2128
No log 1.4857 52 1.4053 0.0 1.4053 1.1855
No log 1.5429 54 1.0037 0.0032 1.0037 1.0018
No log 1.6 56 0.8322 0.0017 0.8322 0.9123
No log 1.6571 58 0.8154 0.0260 0.8154 0.9030
No log 1.7143 60 1.0003 -0.1281 1.0003 1.0002
No log 1.7714 62 0.8736 -0.1257 0.8736 0.9346
No log 1.8286 64 0.7555 0.0 0.7555 0.8692
No log 1.8857 66 0.6819 -0.0626 0.6819 0.8258
No log 1.9429 68 0.8126 -0.0101 0.8126 0.9015
No log 2.0 70 1.1081 0.0100 1.1081 1.0527
No log 2.0571 72 1.6564 0.0425 1.6564 1.2870
No log 2.1143 74 1.2172 -0.0490 1.2172 1.1033
No log 2.1714 76 0.7733 -0.0331 0.7733 0.8794
No log 2.2286 78 0.6986 0.0555 0.6986 0.8358
No log 2.2857 80 0.7065 0.1657 0.7065 0.8405
No log 2.3429 82 0.8723 -0.0079 0.8723 0.9340
No log 2.4 84 1.0073 -0.0236 1.0073 1.0036
No log 2.4571 86 0.9927 -0.0236 0.9927 0.9963
No log 2.5143 88 0.9628 -0.0345 0.9628 0.9812
No log 2.5714 90 1.0109 -0.0704 1.0109 1.0054
No log 2.6286 92 0.9829 -0.1289 0.9829 0.9914
No log 2.6857 94 0.9425 -0.0966 0.9425 0.9708
No log 2.7429 96 0.9446 -0.1285 0.9446 0.9719
No log 2.8 98 0.9625 -0.0423 0.9625 0.9811
No log 2.8571 100 0.9144 -0.0923 0.9144 0.9562
No log 2.9143 102 0.9143 -0.1650 0.9143 0.9562
No log 2.9714 104 0.9081 -0.1671 0.9081 0.9530
No log 3.0286 106 0.9359 -0.0571 0.9359 0.9674
No log 3.0857 108 0.9177 -0.1660 0.9177 0.9580
No log 3.1429 110 0.7934 -0.0679 0.7934 0.8907
No log 3.2 112 0.7385 0.0 0.7385 0.8593
No log 3.2571 114 0.7659 -0.0626 0.7659 0.8751
No log 3.3143 116 0.8933 -0.1676 0.8933 0.9451
No log 3.3714 118 0.8260 -0.1236 0.8260 0.9088
No log 3.4286 120 0.8259 -0.0739 0.8259 0.9088
No log 3.4857 122 0.8203 -0.0753 0.8203 0.9057
No log 3.5429 124 0.8780 -0.1695 0.8780 0.9370
No log 3.6 126 0.9350 -0.0861 0.9350 0.9669
No log 3.6571 128 0.8871 -0.0442 0.8871 0.9418
No log 3.7143 130 0.8008 -0.1730 0.8008 0.8949
No log 3.7714 132 0.7926 -0.0160 0.7926 0.8903
No log 3.8286 134 0.7984 -0.0160 0.7984 0.8935
No log 3.8857 136 0.7638 0.0506 0.7638 0.8740
No log 3.9429 138 0.7965 -0.0035 0.7965 0.8924
No log 4.0 140 0.8210 -0.1074 0.8210 0.9061
No log 4.0571 142 0.8489 -0.1527 0.8489 0.9214
No log 4.1143 144 0.8561 -0.1330 0.8561 0.9253
No log 4.1714 146 0.8832 -0.1146 0.8832 0.9398
No log 4.2286 148 0.8726 -0.1330 0.8726 0.9341
No log 4.2857 150 0.9705 -0.0861 0.9705 0.9851
No log 4.3429 152 1.1682 -0.0595 1.1682 1.0809
No log 4.4 154 0.8786 -0.0091 0.8786 0.9373
No log 4.4571 156 0.8600 -0.0849 0.8600 0.9273
No log 4.5143 158 0.8507 -0.0427 0.8507 0.9224
No log 4.5714 160 0.8733 -0.1601 0.8733 0.9345
No log 4.6286 162 0.9738 -0.1475 0.9738 0.9868
No log 4.6857 164 0.9678 -0.0685 0.9678 0.9838
No log 4.7429 166 1.0046 -0.0650 1.0046 1.0023
No log 4.8 168 1.0393 -0.0855 1.0393 1.0195
No log 4.8571 170 1.0381 -0.0786 1.0381 1.0189
No log 4.9143 172 0.9956 -0.1323 0.9956 0.9978
No log 4.9714 174 0.8902 -0.1204 0.8902 0.9435
No log 5.0286 176 0.8455 -0.0532 0.8455 0.9195
No log 5.0857 178 0.8090 -0.1010 0.8090 0.8995
No log 5.1429 180 0.7982 -0.1551 0.7982 0.8934
No log 5.2 182 0.8399 -0.1833 0.8399 0.9165
No log 5.2571 184 0.9042 -0.1843 0.9042 0.9509
No log 5.3143 186 0.8779 -0.1329 0.8779 0.9370
No log 5.3714 188 0.9532 -0.1530 0.9532 0.9763
No log 5.4286 190 1.0593 -0.0604 1.0593 1.0292
No log 5.4857 192 1.0597 -0.0996 1.0597 1.0294
No log 5.5429 194 1.0323 -0.0721 1.0323 1.0160
No log 5.6 196 0.9318 -0.0934 0.9318 0.9653
No log 5.6571 198 0.8373 -0.0571 0.8373 0.9150
No log 5.7143 200 0.8178 -0.0626 0.8178 0.9043
No log 5.7714 202 0.8378 -0.1473 0.8378 0.9153
No log 5.8286 204 0.8928 -0.0724 0.8928 0.9449
No log 5.8857 206 0.8923 0.0251 0.8923 0.9446
No log 5.9429 208 0.8816 -0.0271 0.8816 0.9389
No log 6.0 210 0.9284 -0.0116 0.9284 0.9635
No log 6.0571 212 1.0347 -0.0204 1.0347 1.0172
No log 6.1143 214 1.0317 -0.1234 1.0317 1.0157
No log 6.1714 216 0.9966 -0.0936 0.9966 0.9983
No log 6.2286 218 0.9159 -0.0566 0.9159 0.9570
No log 6.2857 220 0.9282 -0.0408 0.9282 0.9634
No log 6.3429 222 0.9553 -0.0262 0.9553 0.9774
No log 6.4 224 0.8703 -0.1329 0.8703 0.9329
No log 6.4571 226 0.8374 -0.0881 0.8374 0.9151
No log 6.5143 228 0.8815 -0.1474 0.8815 0.9389
No log 6.5714 230 0.9205 -0.0377 0.9205 0.9594
No log 6.6286 232 0.8961 -0.0934 0.8961 0.9466
No log 6.6857 234 0.9176 -0.0377 0.9176 0.9579
No log 6.7429 236 1.0056 -0.0551 1.0056 1.0028
No log 6.8 238 1.0354 0.0147 1.0354 1.0175
No log 6.8571 240 0.9602 -0.0811 0.9602 0.9799
No log 6.9143 242 0.9450 -0.0741 0.9450 0.9721
No log 6.9714 244 0.9795 -0.0508 0.9795 0.9897
No log 7.0286 246 1.0703 -0.0211 1.0703 1.0345
No log 7.0857 248 0.9483 -0.1647 0.9483 0.9738
No log 7.1429 250 0.8800 -0.0661 0.8800 0.9381
No log 7.2 252 0.8700 -0.0620 0.8700 0.9327
No log 7.2571 254 0.9608 -0.0650 0.9608 0.9802
No log 7.3143 256 1.0760 -0.0337 1.0760 1.0373
No log 7.3714 258 0.9544 -0.0650 0.9544 0.9769
No log 7.4286 260 0.8122 -0.0406 0.8122 0.9012
No log 7.4857 262 0.7728 0.0 0.7728 0.8791
No log 7.5429 264 0.8003 -0.0499 0.8003 0.8946
No log 7.6 266 0.8803 -0.1086 0.8803 0.9382
No log 7.6571 268 0.9095 -0.1354 0.9095 0.9537
No log 7.7143 270 0.9263 -0.0750 0.9263 0.9624
No log 7.7714 272 1.0059 0.0377 1.0059 1.0030
No log 7.8286 274 1.0827 0.0028 1.0827 1.0405
No log 7.8857 276 0.9654 0.0092 0.9654 0.9825
No log 7.9429 278 0.8631 -0.1263 0.8631 0.9290
No log 8.0 280 0.8006 -0.0101 0.8006 0.8948
No log 8.0571 282 0.8241 -0.0309 0.8241 0.9078
No log 8.1143 284 0.7679 0.0035 0.7679 0.8763
No log 8.1714 286 0.7276 -0.0499 0.7276 0.8530
No log 8.2286 288 0.7639 0.0662 0.7639 0.8740
No log 8.2857 290 0.8127 -0.1551 0.8127 0.9015
No log 8.3429 292 0.8945 -0.0573 0.8945 0.9458
No log 8.4 294 0.9381 -0.1464 0.9381 0.9686
No log 8.4571 296 1.1337 -0.0996 1.1337 1.0648
No log 8.5143 298 1.3616 -0.0057 1.3616 1.1669
No log 8.5714 300 1.3270 -0.0057 1.3270 1.1519
No log 8.6286 302 1.1439 -0.1080 1.1439 1.0695
No log 8.6857 304 0.8826 -0.1033 0.8826 0.9395
No log 8.7429 306 0.8101 -0.0065 0.8101 0.9001
No log 8.8 308 0.7884 0.0 0.7884 0.8879
No log 8.8571 310 0.8146 -0.1137 0.8146 0.9026
No log 8.9143 312 0.8526 -0.0774 0.8526 0.9234
No log 8.9714 314 0.9441 -0.0685 0.9441 0.9716
No log 9.0286 316 0.9576 -0.0236 0.9576 0.9786
No log 9.0857 318 0.8489 -0.1678 0.8489 0.9214
No log 9.1429 320 0.8277 0.0479 0.8277 0.9098
No log 9.2 322 0.8114 -0.0591 0.8114 0.9008
No log 9.2571 324 0.8205 0.0229 0.8205 0.9058
No log 9.3143 326 0.8439 0.0310 0.8439 0.9187
No log 9.3714 328 0.8549 -0.0040 0.8549 0.9246
No log 9.4286 330 0.8116 -0.0774 0.8116 0.9009
No log 9.4857 332 0.8125 -0.0881 0.8125 0.9014
No log 9.5429 334 0.9197 -0.0722 0.9197 0.9590
No log 9.6 336 1.0330 -0.0079 1.0330 1.0164
No log 9.6571 338 1.0455 -0.0041 1.0455 1.0225
No log 9.7143 340 0.9565 -0.0650 0.9565 0.9780
No log 9.7714 342 0.9235 -0.0801 0.9235 0.9610
No log 9.8286 344 0.8984 -0.2837 0.8984 0.9478
No log 9.8857 346 0.8853 -0.2580 0.8853 0.9409
No log 9.9429 348 0.9258 -0.0843 0.9258 0.9622
No log 10.0 350 0.9857 -0.1572 0.9857 0.9928
No log 10.0571 352 0.9354 -0.0346 0.9354 0.9672
No log 10.1143 354 0.8137 -0.1067 0.8137 0.9021
No log 10.1714 356 0.7655 -0.0679 0.7655 0.8749
No log 10.2286 358 0.8108 0.0296 0.8108 0.9005
No log 10.2857 360 0.8068 -0.0644 0.8068 0.8982
No log 10.3429 362 0.9305 -0.0377 0.9305 0.9646
No log 10.4 364 1.0850 -0.0786 1.0850 1.0416
No log 10.4571 366 1.0683 -0.0930 1.0683 1.0336
No log 10.5143 368 0.9542 -0.0801 0.9542 0.9768
No log 10.5714 370 0.9032 -0.0934 0.9032 0.9504
No log 10.6286 372 0.9396 -0.1054 0.9396 0.9693
No log 10.6857 374 0.9804 -0.0521 0.9804 0.9902
No log 10.7429 376 0.8791 0.0229 0.8791 0.9376
No log 10.8 378 0.8115 -0.1137 0.8115 0.9008
No log 10.8571 380 0.8102 -0.0188 0.8102 0.9001
No log 10.9143 382 0.8791 -0.0801 0.8791 0.9376
No log 10.9714 384 0.8483 -0.1604 0.8483 0.9210
No log 11.0286 386 0.9284 -0.0016 0.9284 0.9635
No log 11.0857 388 1.0612 -0.0164 1.0612 1.0302
No log 11.1429 390 1.1260 0.0305 1.1260 1.0611
No log 11.2 392 0.9722 -0.0346 0.9722 0.9860
No log 11.2571 394 0.8014 -0.0939 0.8014 0.8952
No log 11.3143 396 0.7866 -0.0609 0.7866 0.8869
No log 11.3714 398 0.7886 -0.0609 0.7886 0.8880
No log 11.4286 400 0.8067 -0.0881 0.8067 0.8982
No log 11.4857 402 0.9507 0.0008 0.9507 0.9751
No log 11.5429 404 1.1028 0.0267 1.1028 1.0502
No log 11.6 406 1.1466 -0.0292 1.1466 1.0708
No log 11.6571 408 1.0339 -0.0187 1.0339 1.0168
No log 11.7143 410 0.9634 -0.1324 0.9634 0.9815
No log 11.7714 412 0.9040 0.0345 0.9040 0.9508
No log 11.8286 414 0.8583 -0.0591 0.8583 0.9265
No log 11.8857 416 0.8774 0.0328 0.8774 0.9367
No log 11.9429 418 0.8805 0.0310 0.8805 0.9384
No log 12.0 420 0.8590 -0.0889 0.8590 0.9268
No log 12.0571 422 0.8497 -0.1659 0.8497 0.9218
No log 12.1143 424 0.8414 -0.0248 0.8414 0.9173
No log 12.1714 426 0.8241 -0.0363 0.8241 0.9078
No log 12.2286 428 0.8140 0.1199 0.8140 0.9022
No log 12.2857 430 0.8006 0.0621 0.8006 0.8948
No log 12.3429 432 0.7898 0.0621 0.7898 0.8887
No log 12.4 434 0.7672 0.0 0.7672 0.8759
No log 12.4571 436 0.7616 0.0555 0.7616 0.8727
No log 12.5143 438 0.7727 0.0460 0.7727 0.8790
No log 12.5714 440 0.7741 0.0555 0.7741 0.8798
No log 12.6286 442 0.7940 0.0621 0.7940 0.8911
No log 12.6857 444 0.8411 -0.0248 0.8411 0.9171
No log 12.7429 446 0.8616 -0.0248 0.8616 0.9282
No log 12.8 448 0.8987 -0.0066 0.8987 0.9480
No log 12.8571 450 0.8935 -0.0066 0.8935 0.9453
No log 12.9143 452 0.8620 -0.0284 0.8620 0.9284
No log 12.9714 454 0.8522 -0.0322 0.8522 0.9231
No log 13.0286 456 0.8546 0.0155 0.8546 0.9244
No log 13.0857 458 0.8719 -0.0406 0.8719 0.9338
No log 13.1429 460 0.8746 0.0181 0.8746 0.9352
No log 13.2 462 0.8682 0.0181 0.8682 0.9318
No log 13.2571 464 0.8325 -0.0406 0.8325 0.9124
No log 13.3143 466 0.8115 -0.1137 0.8115 0.9008
No log 13.3714 468 0.7803 0.0 0.7803 0.8833
No log 13.4286 470 0.7546 0.0 0.7546 0.8687
No log 13.4857 472 0.7485 0.0 0.7485 0.8652
No log 13.5429 474 0.7521 0.0 0.7521 0.8672
No log 13.6 476 0.7854 0.0555 0.7854 0.8862
No log 13.6571 478 0.8003 0.0 0.8003 0.8946
No log 13.7143 480 0.8417 0.0128 0.8417 0.9175
No log 13.7714 482 0.9031 0.0328 0.9031 0.9503
No log 13.8286 484 0.9631 0.0512 0.9631 0.9814
No log 13.8857 486 0.9542 0.1491 0.9542 0.9768
No log 13.9429 488 0.8891 0.0377 0.8891 0.9429
No log 14.0 490 0.8402 -0.0591 0.8402 0.9166
No log 14.0571 492 0.7719 0.0649 0.7719 0.8786
No log 14.1143 494 0.7430 0.0 0.7430 0.8620
No log 14.1714 496 0.7668 0.0621 0.7668 0.8757
No log 14.2286 498 0.8405 -0.0150 0.8405 0.9168
0.3958 14.2857 500 0.9424 -0.0606 0.9424 0.9708
0.3958 14.3429 502 0.9478 -0.1466 0.9478 0.9736
0.3958 14.4 504 0.9632 -0.1413 0.9632 0.9814
0.3958 14.4571 506 1.1464 -0.0410 1.1464 1.0707
0.3958 14.5143 508 1.2364 -0.0292 1.2364 1.1119
0.3958 14.5714 510 1.1540 -0.0360 1.1540 1.0742
0.3958 14.6286 512 1.0094 -0.0289 1.0094 1.0047
0.3958 14.6857 514 0.8993 -0.1470 0.8993 0.9483
0.3958 14.7429 516 0.8238 -0.1468 0.8238 0.9077
0.3958 14.8 518 0.7722 0.0555 0.7722 0.8788
0.3958 14.8571 520 0.7640 0.0555 0.7640 0.8741
0.3958 14.9143 522 0.7744 0.0 0.7744 0.8800
0.3958 14.9714 524 0.8135 0.0206 0.8135 0.9019
0.3958 15.0286 526 0.8924 -0.0016 0.8924 0.9447
0.3958 15.0857 528 0.9577 -0.0377 0.9577 0.9786
0.3958 15.1429 530 0.9967 0.0052 0.9967 0.9984
0.3958 15.2 532 0.9825 -0.0346 0.9825 0.9912
0.3958 15.2571 534 0.9224 -0.0016 0.9224 0.9604
0.3958 15.3143 536 0.8616 0.0251 0.8616 0.9282
0.3958 15.3714 538 0.8082 -0.0406 0.8082 0.8990
0.3958 15.4286 540 0.7972 -0.0551 0.7972 0.8929
0.3958 15.4857 542 0.7972 0.0 0.7972 0.8929
0.3958 15.5429 544 0.8162 0.0 0.8162 0.9034
0.3958 15.6 546 0.8144 0.0 0.8144 0.9024
0.3958 15.6571 548 0.8425 -0.1137 0.8425 0.9179
0.3958 15.7143 550 0.8695 -0.1470 0.8695 0.9325
0.3958 15.7714 552 0.8927 -0.1745 0.8927 0.9448
0.3958 15.8286 554 0.9069 -0.2056 0.9069 0.9523
0.3958 15.8857 556 0.9169 -0.2056 0.9169 0.9576

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k14_task3_organization

Finetuned
(4019)
this model