ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k11_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8851
  • Qwk: 0.4593
  • Mse: 0.8851
  • Rmse: 0.9408

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 4.1532 -0.0217 4.1532 2.0379
No log 0.1111 4 2.4601 0.0396 2.4601 1.5685
No log 0.1667 6 1.6081 -0.0046 1.6081 1.2681
No log 0.2222 8 1.1824 0.1024 1.1824 1.0874
No log 0.2778 10 1.1077 0.3370 1.1077 1.0525
No log 0.3333 12 0.9821 0.2713 0.9821 0.9910
No log 0.3889 14 0.9744 0.2933 0.9744 0.9871
No log 0.4444 16 0.9770 0.3676 0.9770 0.9884
No log 0.5 18 0.9400 0.2765 0.9400 0.9696
No log 0.5556 20 1.0621 0.3979 1.0621 1.0306
No log 0.6111 22 1.0569 0.3564 1.0569 1.0281
No log 0.6667 24 0.9725 0.3243 0.9725 0.9862
No log 0.7222 26 1.0702 0.2135 1.0702 1.0345
No log 0.7778 28 1.0557 0.2288 1.0557 1.0275
No log 0.8333 30 0.9904 0.3117 0.9904 0.9952
No log 0.8889 32 1.3089 0.2870 1.3089 1.1441
No log 0.9444 34 1.5167 0.2123 1.5167 1.2316
No log 1.0 36 1.3492 0.2244 1.3492 1.1615
No log 1.0556 38 1.0962 0.3424 1.0962 1.0470
No log 1.1111 40 1.0138 0.1783 1.0138 1.0069
No log 1.1667 42 1.0681 0.1601 1.0681 1.0335
No log 1.2222 44 1.0659 0.1601 1.0659 1.0324
No log 1.2778 46 1.0287 0.2365 1.0287 1.0143
No log 1.3333 48 1.0881 0.2282 1.0881 1.0431
No log 1.3889 50 1.2336 0.3295 1.2336 1.1107
No log 1.4444 52 1.2116 0.3295 1.2116 1.1007
No log 1.5 54 1.0756 0.2639 1.0756 1.0371
No log 1.5556 56 1.0440 0.2812 1.0440 1.0217
No log 1.6111 58 1.0558 0.2933 1.0558 1.0275
No log 1.6667 60 1.0669 0.2545 1.0669 1.0329
No log 1.7222 62 1.0754 0.2880 1.0754 1.0370
No log 1.7778 64 1.0586 0.3488 1.0586 1.0289
No log 1.8333 66 1.0230 0.3221 1.0230 1.0114
No log 1.8889 68 1.0158 0.1881 1.0158 1.0079
No log 1.9444 70 1.0205 0.2154 1.0205 1.0102
No log 2.0 72 1.0484 0.2721 1.0484 1.0239
No log 2.0556 74 1.0300 0.3128 1.0300 1.0149
No log 2.1111 76 1.1190 0.2904 1.1190 1.0578
No log 2.1667 78 1.1564 0.2527 1.1564 1.0754
No log 2.2222 80 0.9989 0.3485 0.9989 0.9995
No log 2.2778 82 0.9325 0.3276 0.9325 0.9657
No log 2.3333 84 0.8924 0.3784 0.8924 0.9446
No log 2.3889 86 0.8973 0.4126 0.8973 0.9472
No log 2.4444 88 0.9166 0.3502 0.9166 0.9574
No log 2.5 90 0.8872 0.3661 0.8872 0.9419
No log 2.5556 92 0.9908 0.4176 0.9908 0.9954
No log 2.6111 94 1.1738 0.4181 1.1738 1.0834
No log 2.6667 96 1.3270 0.3629 1.3270 1.1519
No log 2.7222 98 1.3189 0.3707 1.3189 1.1485
No log 2.7778 100 1.3835 0.3350 1.3835 1.1762
No log 2.8333 102 1.4765 0.2831 1.4765 1.2151
No log 2.8889 104 1.3412 0.2981 1.3412 1.1581
No log 2.9444 106 1.0778 0.4334 1.0778 1.0381
No log 3.0 108 0.9955 0.4240 0.9955 0.9977
No log 3.0556 110 0.9780 0.4157 0.9780 0.9890
No log 3.1111 112 0.9075 0.4484 0.9075 0.9526
No log 3.1667 114 1.0147 0.4610 1.0147 1.0073
No log 3.2222 116 1.0936 0.4490 1.0936 1.0458
No log 3.2778 118 0.9898 0.4818 0.9898 0.9949
No log 3.3333 120 0.8865 0.3996 0.8865 0.9416
No log 3.3889 122 0.9173 0.4644 0.9173 0.9578
No log 3.4444 124 0.9177 0.5060 0.9177 0.9580
No log 3.5 126 0.9592 0.5416 0.9592 0.9794
No log 3.5556 128 1.0475 0.4894 1.0475 1.0235
No log 3.6111 130 1.0910 0.4701 1.0910 1.0445
No log 3.6667 132 1.0910 0.4941 1.0910 1.0445
No log 3.7222 134 1.0762 0.4839 1.0762 1.0374
No log 3.7778 136 0.9908 0.5495 0.9908 0.9954
No log 3.8333 138 0.9784 0.4870 0.9784 0.9891
No log 3.8889 140 0.9636 0.4538 0.9636 0.9816
No log 3.9444 142 0.8755 0.4863 0.8755 0.9357
No log 4.0 144 0.8210 0.4778 0.8210 0.9061
No log 4.0556 146 0.7985 0.4703 0.7985 0.8936
No log 4.1111 148 0.7894 0.5021 0.7894 0.8885
No log 4.1667 150 0.8329 0.4976 0.8329 0.9126
No log 4.2222 152 0.8378 0.5691 0.8378 0.9153
No log 4.2778 154 0.8587 0.5510 0.8587 0.9267
No log 4.3333 156 0.9245 0.5534 0.9245 0.9615
No log 4.3889 158 0.8825 0.5403 0.8825 0.9394
No log 4.4444 160 0.9236 0.5627 0.9236 0.9610
No log 4.5 162 1.0485 0.4461 1.0485 1.0240
No log 4.5556 164 1.0711 0.4159 1.0711 1.0349
No log 4.6111 166 1.0341 0.4213 1.0341 1.0169
No log 4.6667 168 1.1024 0.4014 1.1024 1.0499
No log 4.7222 170 1.2421 0.3453 1.2421 1.1145
No log 4.7778 172 1.2675 0.3803 1.2675 1.1258
No log 4.8333 174 1.1801 0.3935 1.1801 1.0863
No log 4.8889 176 1.0570 0.5059 1.0570 1.0281
No log 4.9444 178 0.9916 0.5431 0.9916 0.9958
No log 5.0 180 1.0461 0.4931 1.0461 1.0228
No log 5.0556 182 1.2090 0.4256 1.2090 1.0996
No log 5.1111 184 1.1168 0.4989 1.1168 1.0568
No log 5.1667 186 0.9255 0.5087 0.9255 0.9620
No log 5.2222 188 1.0801 0.4772 1.0801 1.0393
No log 5.2778 190 1.0731 0.4681 1.0731 1.0359
No log 5.3333 192 0.8676 0.5555 0.8676 0.9315
No log 5.3889 194 0.8932 0.4949 0.8932 0.9451
No log 5.4444 196 0.9892 0.4209 0.9892 0.9946
No log 5.5 198 0.9430 0.3929 0.9430 0.9711
No log 5.5556 200 0.8761 0.4898 0.8761 0.9360
No log 5.6111 202 0.8840 0.5264 0.8840 0.9402
No log 5.6667 204 0.8745 0.5053 0.8745 0.9352
No log 5.7222 206 0.8615 0.5291 0.8615 0.9282
No log 5.7778 208 0.9178 0.4757 0.9178 0.9580
No log 5.8333 210 0.9397 0.5325 0.9397 0.9694
No log 5.8889 212 0.9735 0.5040 0.9735 0.9867
No log 5.9444 214 1.0241 0.4738 1.0241 1.0120
No log 6.0 216 0.9189 0.4839 0.9189 0.9586
No log 6.0556 218 0.8630 0.5291 0.8630 0.9290
No log 6.1111 220 0.8351 0.5291 0.8351 0.9138
No log 6.1667 222 0.7958 0.5364 0.7958 0.8921
No log 6.2222 224 0.7811 0.5044 0.7811 0.8838
No log 6.2778 226 0.7929 0.4378 0.7929 0.8904
No log 6.3333 228 0.7986 0.4903 0.7986 0.8937
No log 6.3889 230 0.8451 0.5038 0.8451 0.9193
No log 6.4444 232 0.9030 0.4649 0.9030 0.9503
No log 6.5 234 1.0792 0.3658 1.0792 1.0388
No log 6.5556 236 1.1583 0.3737 1.1583 1.0762
No log 6.6111 238 1.1025 0.3646 1.1025 1.0500
No log 6.6667 240 1.0201 0.4281 1.0201 1.0100
No log 6.7222 242 0.9548 0.4393 0.9548 0.9771
No log 6.7778 244 0.9337 0.4518 0.9337 0.9663
No log 6.8333 246 1.0048 0.3913 1.0048 1.0024
No log 6.8889 248 0.9724 0.4103 0.9724 0.9861
No log 6.9444 250 0.8783 0.4228 0.8783 0.9372
No log 7.0 252 0.8487 0.4268 0.8487 0.9213
No log 7.0556 254 0.8871 0.3712 0.8871 0.9418
No log 7.1111 256 0.9446 0.4255 0.9446 0.9719
No log 7.1667 258 1.0129 0.4472 1.0129 1.0064
No log 7.2222 260 0.9752 0.4167 0.9752 0.9875
No log 7.2778 262 0.8787 0.4859 0.8787 0.9374
No log 7.3333 264 0.9052 0.3945 0.9052 0.9514
No log 7.3889 266 0.9792 0.4222 0.9792 0.9896
No log 7.4444 268 0.9728 0.4222 0.9728 0.9863
No log 7.5 270 0.9281 0.4107 0.9281 0.9634
No log 7.5556 272 0.9626 0.3985 0.9626 0.9811
No log 7.6111 274 0.8956 0.4663 0.8956 0.9463
No log 7.6667 276 0.8896 0.4676 0.8896 0.9432
No log 7.7222 278 0.9557 0.4327 0.9557 0.9776
No log 7.7778 280 0.9850 0.4228 0.9850 0.9925
No log 7.8333 282 1.1222 0.4471 1.1222 1.0593
No log 7.8889 284 1.0461 0.4232 1.0461 1.0228
No log 7.9444 286 0.8866 0.4321 0.8866 0.9416
No log 8.0 288 0.8675 0.4329 0.8675 0.9314
No log 8.0556 290 0.8919 0.2791 0.8919 0.9444
No log 8.1111 292 0.9495 0.3176 0.9495 0.9744
No log 8.1667 294 1.0281 0.4565 1.0281 1.0139
No log 8.2222 296 1.0370 0.4186 1.0370 1.0183
No log 8.2778 298 0.9791 0.4458 0.9791 0.9895
No log 8.3333 300 1.0087 0.4033 1.0087 1.0043
No log 8.3889 302 1.1001 0.4681 1.1001 1.0488
No log 8.4444 304 1.0831 0.4454 1.0831 1.0407
No log 8.5 306 0.9369 0.3590 0.9369 0.9680
No log 8.5556 308 0.8315 0.3779 0.8315 0.9119
No log 8.6111 310 0.9113 0.3814 0.9113 0.9546
No log 8.6667 312 0.9225 0.4315 0.9225 0.9605
No log 8.7222 314 0.8328 0.3985 0.8328 0.9126
No log 8.7778 316 0.8079 0.5237 0.8079 0.8988
No log 8.8333 318 0.8263 0.4996 0.8263 0.9090
No log 8.8889 320 0.8429 0.5127 0.8429 0.9181
No log 8.9444 322 0.8593 0.5040 0.8593 0.9270
No log 9.0 324 0.8631 0.5302 0.8631 0.9291
No log 9.0556 326 0.9552 0.4834 0.9552 0.9773
No log 9.1111 328 0.9234 0.4834 0.9234 0.9609
No log 9.1667 330 0.8346 0.4690 0.8346 0.9136
No log 9.2222 332 0.8266 0.4805 0.8266 0.9092
No log 9.2778 334 0.8547 0.4567 0.8547 0.9245
No log 9.3333 336 1.0023 0.4890 1.0023 1.0012
No log 9.3889 338 1.0183 0.5079 1.0183 1.0091
No log 9.4444 340 0.9397 0.4828 0.9397 0.9694
No log 9.5 342 0.8969 0.4554 0.8969 0.9470
No log 9.5556 344 0.8811 0.5502 0.8811 0.9387
No log 9.6111 346 0.9338 0.4796 0.9338 0.9663
No log 9.6667 348 1.0379 0.4417 1.0379 1.0188
No log 9.7222 350 1.1468 0.4152 1.1468 1.0709
No log 9.7778 352 1.1036 0.4341 1.1036 1.0505
No log 9.8333 354 0.9648 0.3914 0.9648 0.9822
No log 9.8889 356 0.8926 0.4030 0.8926 0.9448
No log 9.9444 358 0.8756 0.3933 0.8756 0.9357
No log 10.0 360 0.8530 0.4198 0.8530 0.9236
No log 10.0556 362 0.8571 0.3879 0.8571 0.9258
No log 10.1111 364 0.8908 0.3250 0.8908 0.9438
No log 10.1667 366 0.8864 0.4321 0.8864 0.9415
No log 10.2222 368 0.8915 0.4558 0.8915 0.9442
No log 10.2778 370 0.9442 0.4468 0.9442 0.9717
No log 10.3333 372 0.9961 0.4472 0.9961 0.9980
No log 10.3889 374 1.0451 0.4479 1.0451 1.0223
No log 10.4444 376 1.0482 0.4701 1.0482 1.0238
No log 10.5 378 1.0141 0.4726 1.0141 1.0070
No log 10.5556 380 0.9754 0.5019 0.9754 0.9876
No log 10.6111 382 0.9472 0.5383 0.9472 0.9732
No log 10.6667 384 0.9646 0.4862 0.9646 0.9821
No log 10.7222 386 0.9510 0.5083 0.9510 0.9752
No log 10.7778 388 0.9367 0.4453 0.9367 0.9678
No log 10.8333 390 0.9297 0.4552 0.9297 0.9642
No log 10.8889 392 0.9101 0.4727 0.9101 0.9540
No log 10.9444 394 0.9155 0.5208 0.9155 0.9568
No log 11.0 396 0.9569 0.5066 0.9569 0.9782
No log 11.0556 398 1.0881 0.4493 1.0881 1.0431
No log 11.1111 400 1.0901 0.4249 1.0901 1.0441
No log 11.1667 402 0.9702 0.4940 0.9702 0.9850
No log 11.2222 404 0.9139 0.4872 0.9139 0.9560
No log 11.2778 406 0.8896 0.5203 0.8896 0.9432
No log 11.3333 408 0.8783 0.4662 0.8783 0.9372
No log 11.3889 410 0.8766 0.4661 0.8766 0.9363
No log 11.4444 412 0.8941 0.4321 0.8941 0.9455
No log 11.5 414 0.9116 0.4650 0.9116 0.9548
No log 11.5556 416 0.9234 0.5399 0.9234 0.9609
No log 11.6111 418 0.9745 0.4837 0.9745 0.9872
No log 11.6667 420 1.0306 0.4463 1.0306 1.0152
No log 11.7222 422 0.9778 0.4241 0.9778 0.9889
No log 11.7778 424 0.8368 0.5199 0.8368 0.9148
No log 11.8333 426 0.7997 0.5124 0.7997 0.8943
No log 11.8889 428 0.7984 0.5490 0.7984 0.8935
No log 11.9444 430 0.8009 0.5232 0.8009 0.8949
No log 12.0 432 0.7840 0.5701 0.7840 0.8855
No log 12.0556 434 0.7800 0.5368 0.7800 0.8832
No log 12.1111 436 0.7748 0.5680 0.7748 0.8802
No log 12.1667 438 0.8010 0.5304 0.8010 0.8950
No log 12.2222 440 0.8258 0.5255 0.8258 0.9088
No log 12.2778 442 0.8224 0.5255 0.8224 0.9068
No log 12.3333 444 0.8169 0.5368 0.8169 0.9038
No log 12.3889 446 0.8439 0.4911 0.8439 0.9187
No log 12.4444 448 0.8002 0.6018 0.8002 0.8945
No log 12.5 450 0.7853 0.5892 0.7853 0.8862
No log 12.5556 452 0.8072 0.5895 0.8072 0.8984
No log 12.6111 454 0.8126 0.5679 0.8126 0.9014
No log 12.6667 456 0.8424 0.4911 0.8424 0.9178
No log 12.7222 458 0.8355 0.4579 0.8355 0.9141
No log 12.7778 460 0.7802 0.4603 0.7802 0.8833
No log 12.8333 462 0.7567 0.5690 0.7567 0.8699
No log 12.8889 464 0.7585 0.5383 0.7585 0.8709
No log 12.9444 466 0.7578 0.5383 0.7578 0.8705
No log 13.0 468 0.7543 0.5383 0.7543 0.8685
No log 13.0556 470 0.7662 0.5383 0.7662 0.8753
No log 13.1111 472 0.8100 0.4840 0.8100 0.9000
No log 13.1667 474 0.8417 0.5037 0.8417 0.9174
No log 13.2222 476 0.8022 0.5291 0.8022 0.8957
No log 13.2778 478 0.7820 0.5818 0.7820 0.8843
No log 13.3333 480 0.7770 0.5972 0.7770 0.8815
No log 13.3889 482 0.7731 0.5797 0.7731 0.8792
No log 13.4444 484 0.7732 0.5797 0.7732 0.8793
No log 13.5 486 0.7813 0.5797 0.7813 0.8839
No log 13.5556 488 0.7928 0.5489 0.7928 0.8904
No log 13.6111 490 0.8006 0.5378 0.8006 0.8947
No log 13.6667 492 0.8205 0.5267 0.8205 0.9058
No log 13.7222 494 0.8463 0.5159 0.8463 0.9199
No log 13.7778 496 0.8828 0.5159 0.8828 0.9396
No log 13.8333 498 0.8861 0.4856 0.8861 0.9414
0.297 13.8889 500 0.8741 0.5057 0.8741 0.9349
0.297 13.9444 502 0.8794 0.4607 0.8794 0.9377
0.297 14.0 504 0.8709 0.4386 0.8709 0.9332
0.297 14.0556 506 0.8253 0.4037 0.8253 0.9085
0.297 14.1111 508 0.7826 0.5261 0.7826 0.8846
0.297 14.1667 510 0.7908 0.5483 0.7908 0.8892
0.297 14.2222 512 0.7989 0.5424 0.7989 0.8938
0.297 14.2778 514 0.8271 0.4965 0.8271 0.9094
0.297 14.3333 516 0.8839 0.4923 0.8839 0.9402
0.297 14.3889 518 0.8806 0.4930 0.8806 0.9384
0.297 14.4444 520 0.8271 0.4774 0.8271 0.9094
0.297 14.5 522 0.8128 0.5094 0.8128 0.9016
0.297 14.5556 524 0.8241 0.4663 0.8241 0.9078
0.297 14.6111 526 0.8363 0.5074 0.8363 0.9145
0.297 14.6667 528 0.8802 0.4853 0.8802 0.9382
0.297 14.7222 530 0.9279 0.4923 0.9279 0.9633
0.297 14.7778 532 0.9110 0.4928 0.9110 0.9545
0.297 14.8333 534 0.8498 0.4663 0.8498 0.9218
0.297 14.8889 536 0.8279 0.5142 0.8279 0.9099
0.297 14.9444 538 0.8276 0.4676 0.8276 0.9097
0.297 15.0 540 0.8680 0.4499 0.8680 0.9317
0.297 15.0556 542 0.9508 0.4579 0.9508 0.9751
0.297 15.1111 544 0.9454 0.4341 0.9454 0.9723
0.297 15.1667 546 0.8851 0.4593 0.8851 0.9408

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k11_task5_organization

Finetuned
(4019)
this model