ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k15_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7562
  • Qwk: 0.7
  • Mse: 0.7562
  • Rmse: 0.8696

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0286 2 6.9464 0.0176 6.9464 2.6356
No log 0.0571 4 4.7143 0.0735 4.7143 2.1712
No log 0.0857 6 3.5718 0.0 3.5718 1.8899
No log 0.1143 8 3.5853 -0.0104 3.5853 1.8935
No log 0.1429 10 2.5543 0.0147 2.5543 1.5982
No log 0.1714 12 1.8319 0.2655 1.8319 1.3535
No log 0.2 14 1.8718 0.2037 1.8718 1.3681
No log 0.2286 16 1.7988 0.1887 1.7988 1.3412
No log 0.2571 18 1.9845 0.1880 1.9845 1.4087
No log 0.2857 20 2.2135 0.0923 2.2135 1.4878
No log 0.3143 22 2.8884 0.0258 2.8884 1.6995
No log 0.3429 24 3.8149 0.0287 3.8149 1.9532
No log 0.3714 26 3.5540 0.1143 3.5540 1.8852
No log 0.4 28 3.6953 0.1096 3.6953 1.9223
No log 0.4286 30 3.7935 0.0625 3.7935 1.9477
No log 0.4571 32 3.0565 0.1657 3.0565 1.7483
No log 0.4857 34 2.1757 0.2416 2.1757 1.4750
No log 0.5143 36 1.8412 0.2647 1.8412 1.3569
No log 0.5429 38 1.5849 0.2969 1.5849 1.2589
No log 0.5714 40 1.6420 0.3143 1.6420 1.2814
No log 0.6 42 1.9839 0.3356 1.9839 1.4085
No log 0.6286 44 2.4413 0.1687 2.4413 1.5625
No log 0.6571 46 2.4002 0.1893 2.4002 1.5493
No log 0.6857 48 1.7933 0.4161 1.7933 1.3392
No log 0.7143 50 1.2760 0.5075 1.2760 1.1296
No log 0.7429 52 1.1904 0.5649 1.1904 1.0910
No log 0.7714 54 1.3012 0.4848 1.3012 1.1407
No log 0.8 56 1.7278 0.4133 1.7278 1.3144
No log 0.8286 58 1.9667 0.3484 1.9667 1.4024
No log 0.8571 60 1.7595 0.3733 1.7595 1.3265
No log 0.8857 62 1.4079 0.4225 1.4079 1.1865
No log 0.9143 64 1.3175 0.4748 1.3175 1.1478
No log 0.9429 66 1.2932 0.5333 1.2932 1.1372
No log 0.9714 68 1.3365 0.4965 1.3365 1.1561
No log 1.0 70 1.3931 0.5 1.3931 1.1803
No log 1.0286 72 1.2367 0.5793 1.2367 1.1120
No log 1.0571 74 1.2205 0.5714 1.2205 1.1048
No log 1.0857 76 1.4107 0.5399 1.4107 1.1877
No log 1.1143 78 1.2933 0.5679 1.2933 1.1372
No log 1.1429 80 1.1917 0.5 1.1917 1.0917
No log 1.1714 82 1.2596 0.5241 1.2596 1.1223
No log 1.2 84 1.3279 0.5333 1.3279 1.1523
No log 1.2286 86 1.6482 0.5060 1.6482 1.2838
No log 1.2571 88 2.1020 0.3455 2.1020 1.4498
No log 1.2857 90 2.2027 0.3505 2.2027 1.4842
No log 1.3143 92 1.9225 0.3860 1.9225 1.3865
No log 1.3429 94 1.4667 0.5175 1.4667 1.2111
No log 1.3714 96 1.2821 0.4355 1.2821 1.1323
No log 1.4 98 1.2495 0.512 1.2495 1.1178
No log 1.4286 100 1.1542 0.5238 1.1542 1.0743
No log 1.4571 102 1.1721 0.5942 1.1721 1.0826
No log 1.4857 104 1.3017 0.5417 1.3017 1.1409
No log 1.5143 106 1.2649 0.5298 1.2649 1.1247
No log 1.5429 108 1.0557 0.5753 1.0557 1.0275
No log 1.5714 110 0.9854 0.6438 0.9854 0.9927
No log 1.6 112 0.9816 0.6438 0.9816 0.9907
No log 1.6286 114 1.0243 0.6104 1.0243 1.0121
No log 1.6571 116 1.2458 0.5749 1.2458 1.1162
No log 1.6857 118 1.3709 0.5683 1.3709 1.1709
No log 1.7143 120 1.1874 0.6364 1.1874 1.0897
No log 1.7429 122 0.9551 0.6709 0.9551 0.9773
No log 1.7714 124 1.1196 0.6056 1.1196 1.0581
No log 1.8 126 1.1781 0.5362 1.1781 1.0854
No log 1.8286 128 1.0238 0.6621 1.0238 1.0118
No log 1.8571 130 0.9918 0.6575 0.9918 0.9959
No log 1.8857 132 1.3315 0.5783 1.3315 1.1539
No log 1.9143 134 1.5109 0.5556 1.5109 1.2292
No log 1.9429 136 1.4110 0.5106 1.4110 1.1879
No log 1.9714 138 1.1416 0.5468 1.1416 1.0684
No log 2.0 140 1.0308 0.5899 1.0308 1.0153
No log 2.0286 142 0.9228 0.6143 0.9228 0.9606
No log 2.0571 144 0.8597 0.6853 0.8597 0.9272
No log 2.0857 146 0.8359 0.6939 0.8359 0.9143
No log 2.1143 148 0.8297 0.7067 0.8297 0.9109
No log 2.1429 150 0.8794 0.7215 0.8794 0.9378
No log 2.1714 152 0.8922 0.7006 0.8922 0.9445
No log 2.2 154 0.8907 0.7143 0.8907 0.9438
No log 2.2286 156 0.9784 0.6623 0.9784 0.9891
No log 2.2571 158 1.1065 0.6104 1.1065 1.0519
No log 2.2857 160 1.1179 0.5960 1.1179 1.0573
No log 2.3143 162 1.0609 0.5652 1.0609 1.0300
No log 2.3429 164 1.0893 0.5652 1.0893 1.0437
No log 2.3714 166 1.0169 0.6143 1.0169 1.0084
No log 2.4 168 1.1062 0.5912 1.1062 1.0518
No log 2.4286 170 1.3252 0.5862 1.3252 1.1512
No log 2.4571 172 1.3884 0.5862 1.3884 1.1783
No log 2.4857 174 1.2235 0.6071 1.2235 1.1061
No log 2.5143 176 0.9464 0.6479 0.9464 0.9728
No log 2.5429 178 0.8262 0.6993 0.8262 0.9089
No log 2.5714 180 0.8060 0.6944 0.8060 0.8978
No log 2.6 182 0.8627 0.6849 0.8627 0.9288
No log 2.6286 184 0.8031 0.7383 0.8031 0.8962
No log 2.6571 186 0.7903 0.7517 0.7903 0.8890
No log 2.6857 188 0.8012 0.7297 0.8012 0.8951
No log 2.7143 190 0.8231 0.6986 0.8231 0.9073
No log 2.7429 192 0.8520 0.6438 0.8520 0.9231
No log 2.7714 194 0.8145 0.7152 0.8145 0.9025
No log 2.8 196 0.8122 0.7248 0.8122 0.9012
No log 2.8286 198 0.8623 0.7215 0.8623 0.9286
No log 2.8571 200 1.0081 0.6832 1.0081 1.0041
No log 2.8857 202 1.2976 0.5917 1.2976 1.1391
No log 2.9143 204 1.2784 0.5917 1.2784 1.1307
No log 2.9429 206 1.0395 0.6220 1.0395 1.0195
No log 2.9714 208 0.8838 0.6711 0.8838 0.9401
No log 3.0 210 0.8524 0.6944 0.8524 0.9233
No log 3.0286 212 0.9406 0.6377 0.9406 0.9699
No log 3.0571 214 1.0100 0.6377 1.0100 1.0050
No log 3.0857 216 1.0127 0.6119 1.0127 1.0063
No log 3.1143 218 1.0115 0.6119 1.0115 1.0058
No log 3.1429 220 0.9544 0.6119 0.9544 0.9769
No log 3.1714 222 0.9267 0.6187 0.9267 0.9627
No log 3.2 224 0.9030 0.6479 0.9030 0.9503
No log 3.2286 226 0.8839 0.7034 0.8839 0.9401
No log 3.2571 228 0.9672 0.6871 0.9672 0.9835
No log 3.2857 230 1.1859 0.5714 1.1859 1.0890
No log 3.3143 232 1.3141 0.5618 1.3141 1.1463
No log 3.3429 234 1.1716 0.6243 1.1716 1.0824
No log 3.3714 236 1.0515 0.6545 1.0515 1.0254
No log 3.4 238 0.9412 0.6914 0.9412 0.9701
No log 3.4286 240 0.9117 0.7226 0.9117 0.9548
No log 3.4571 242 0.8621 0.7260 0.8621 0.9285
No log 3.4857 244 0.8627 0.7172 0.8627 0.9288
No log 3.5143 246 0.8272 0.7083 0.8272 0.9095
No log 3.5429 248 0.8132 0.7534 0.8132 0.9018
No log 3.5714 250 0.8514 0.6759 0.8514 0.9227
No log 3.6 252 0.9752 0.6752 0.9752 0.9875
No log 3.6286 254 0.9409 0.6790 0.9409 0.9700
No log 3.6571 256 0.8795 0.7081 0.8795 0.9378
No log 3.6857 258 0.8203 0.7342 0.8203 0.9057
No log 3.7143 260 0.7762 0.7467 0.7762 0.8810
No log 3.7429 262 0.7259 0.7550 0.7259 0.8520
No log 3.7714 264 0.7361 0.7467 0.7361 0.8580
No log 3.8 266 0.7981 0.7607 0.7981 0.8934
No log 3.8286 268 0.8051 0.7636 0.8051 0.8972
No log 3.8571 270 0.7144 0.7347 0.7144 0.8452
No log 3.8857 272 0.7257 0.7027 0.7257 0.8519
No log 3.9143 274 0.7388 0.7152 0.7388 0.8595
No log 3.9429 276 0.7179 0.7483 0.7179 0.8473
No log 3.9714 278 0.7587 0.7361 0.7587 0.8711
No log 4.0 280 0.8236 0.75 0.8236 0.9075
No log 4.0286 282 0.8685 0.7383 0.8685 0.9319
No log 4.0571 284 0.8836 0.7133 0.8836 0.9400
No log 4.0857 286 0.8953 0.6573 0.8953 0.9462
No log 4.1143 288 0.8923 0.6429 0.8923 0.9446
No log 4.1429 290 0.9059 0.6620 0.9059 0.9518
No log 4.1714 292 0.8695 0.6620 0.8695 0.9324
No log 4.2 294 0.8667 0.6667 0.8667 0.9310
No log 4.2286 296 1.1244 0.5606 1.1244 1.0604
No log 4.2571 298 1.5355 0.3622 1.5355 1.2391
No log 4.2857 300 1.4296 0.4186 1.4296 1.1957
No log 4.3143 302 0.9776 0.5693 0.9776 0.9887
No log 4.3429 304 0.7531 0.7050 0.7531 0.8678
No log 4.3714 306 0.8028 0.6906 0.8028 0.8960
No log 4.4 308 0.8565 0.6715 0.8565 0.9255
No log 4.4286 310 0.8815 0.6316 0.8815 0.9389
No log 4.4571 312 0.9247 0.6567 0.9247 0.9616
No log 4.4857 314 0.9254 0.6377 0.9254 0.9620
No log 4.5143 316 0.8649 0.6471 0.8649 0.9300
No log 4.5429 318 0.8766 0.6619 0.8766 0.9363
No log 4.5714 320 0.9745 0.6667 0.9745 0.9872
No log 4.6 322 1.0208 0.6104 1.0208 1.0104
No log 4.6286 324 0.9153 0.6968 0.9153 0.9567
No log 4.6571 326 0.8601 0.7205 0.8601 0.9274
No log 4.6857 328 0.8801 0.6709 0.8801 0.9382
No log 4.7143 330 0.8890 0.6667 0.8890 0.9429
No log 4.7429 332 0.9414 0.6443 0.9414 0.9703
No log 4.7714 334 0.8936 0.6667 0.8936 0.9453
No log 4.8 336 0.8397 0.6475 0.8397 0.9164
No log 4.8286 338 0.8259 0.6131 0.8259 0.9088
No log 4.8571 340 0.8363 0.6475 0.8363 0.9145
No log 4.8857 342 0.8868 0.6757 0.8868 0.9417
No log 4.9143 344 0.9504 0.6443 0.9504 0.9749
No log 4.9429 346 0.9175 0.6207 0.9175 0.9579
No log 4.9714 348 0.8378 0.6525 0.8378 0.9153
No log 5.0 350 0.8026 0.6471 0.8026 0.8959
No log 5.0286 352 0.8157 0.6957 0.8157 0.9031
No log 5.0571 354 0.8134 0.6571 0.8134 0.9019
No log 5.0857 356 0.8681 0.6757 0.8681 0.9317
No log 5.1143 358 0.9368 0.6531 0.9368 0.9679
No log 5.1429 360 0.9542 0.6207 0.9542 0.9769
No log 5.1714 362 0.9378 0.6056 0.9378 0.9684
No log 5.2 364 0.8822 0.6232 0.8822 0.9393
No log 5.2286 366 0.8308 0.6944 0.8308 0.9115
No log 5.2571 368 0.8329 0.7075 0.8329 0.9126
No log 5.2857 370 0.8732 0.6483 0.8732 0.9344
No log 5.3143 372 0.9436 0.6667 0.9436 0.9714
No log 5.3429 374 0.9491 0.6667 0.9491 0.9742
No log 5.3714 376 0.8509 0.6939 0.8509 0.9224
No log 5.4 378 0.7951 0.6939 0.7951 0.8917
No log 5.4286 380 0.7591 0.7211 0.7591 0.8713
No log 5.4571 382 0.7716 0.7211 0.7716 0.8784
No log 5.4857 384 0.7950 0.7237 0.7950 0.8916
No log 5.5143 386 0.8681 0.7059 0.8681 0.9317
No log 5.5429 388 0.9203 0.6883 0.9203 0.9593
No log 5.5714 390 0.8971 0.7067 0.8971 0.9472
No log 5.6 392 0.8752 0.6620 0.8752 0.9355
No log 5.6286 394 0.8168 0.7034 0.8168 0.9038
No log 5.6571 396 0.8103 0.7172 0.8103 0.9002
No log 5.6857 398 0.8607 0.7027 0.8607 0.9277
No log 5.7143 400 0.8193 0.7248 0.8193 0.9052
No log 5.7429 402 0.7707 0.7248 0.7707 0.8779
No log 5.7714 404 0.7224 0.75 0.7224 0.8499
No log 5.8 406 0.7005 0.7467 0.7005 0.8370
No log 5.8286 408 0.7263 0.7347 0.7263 0.8522
No log 5.8571 410 0.7484 0.7347 0.7484 0.8651
No log 5.8857 412 0.7573 0.7310 0.7573 0.8702
No log 5.9143 414 0.8368 0.7034 0.8368 0.9148
No log 5.9429 416 0.9208 0.6621 0.9208 0.9596
No log 5.9714 418 0.9464 0.6241 0.9464 0.9728
No log 6.0 420 0.9020 0.7042 0.9020 0.9497
No log 6.0286 422 0.8889 0.6667 0.8889 0.9428
No log 6.0571 424 0.8762 0.6957 0.8762 0.9361
No log 6.0857 426 0.8070 0.6950 0.8070 0.8983
No log 6.1143 428 0.7818 0.7083 0.7818 0.8842
No log 6.1429 430 0.7934 0.6950 0.7934 0.8907
No log 6.1714 432 0.8086 0.6950 0.8086 0.8992
No log 6.2 434 0.8569 0.6897 0.8569 0.9257
No log 6.2286 436 1.0034 0.6301 1.0034 1.0017
No log 6.2571 438 1.0508 0.6259 1.0508 1.0251
No log 6.2857 440 0.9277 0.6892 0.9277 0.9631
No log 6.3143 442 0.7929 0.6901 0.7929 0.8905
No log 6.3429 444 0.7443 0.6857 0.7443 0.8627
No log 6.3714 446 0.7111 0.7172 0.7111 0.8433
No log 6.4 448 0.7368 0.7297 0.7368 0.8584
No log 6.4286 450 0.7673 0.6986 0.7673 0.8760
No log 6.4571 452 0.7413 0.7211 0.7413 0.8610
No log 6.4857 454 0.7660 0.7133 0.7660 0.8752
No log 6.5143 456 0.7893 0.6857 0.7893 0.8884
No log 6.5429 458 0.7908 0.7092 0.7908 0.8893
No log 6.5714 460 0.7976 0.7246 0.7976 0.8931
No log 6.6 462 0.8166 0.7007 0.8166 0.9037
No log 6.6286 464 0.8023 0.6906 0.8023 0.8957
No log 6.6571 466 0.7770 0.6993 0.7770 0.8815
No log 6.6857 468 0.8160 0.7248 0.8160 0.9033
No log 6.7143 470 0.8389 0.7848 0.8389 0.9159
No log 6.7429 472 0.7567 0.7451 0.7567 0.8699
No log 6.7714 474 0.6885 0.7333 0.6885 0.8298
No log 6.8 476 0.6985 0.7712 0.6985 0.8357
No log 6.8286 478 0.6976 0.7550 0.6976 0.8352
No log 6.8571 480 0.6973 0.7532 0.6973 0.8350
No log 6.8857 482 0.7542 0.7532 0.7542 0.8684
No log 6.9143 484 0.8610 0.7582 0.8610 0.9279
No log 6.9429 486 0.8868 0.7285 0.8868 0.9417
No log 6.9714 488 0.8483 0.7368 0.8483 0.9210
No log 7.0 490 0.7904 0.7034 0.7904 0.8890
No log 7.0286 492 0.7520 0.6901 0.7520 0.8672
No log 7.0571 494 0.7368 0.7133 0.7368 0.8584
No log 7.0857 496 0.7581 0.7417 0.7581 0.8707
No log 7.1143 498 0.8295 0.7125 0.8295 0.9108
0.429 7.1429 500 0.8162 0.7125 0.8162 0.9034
0.429 7.1714 502 0.7697 0.7237 0.7697 0.8773
0.429 7.2 504 0.7532 0.7234 0.7532 0.8679
0.429 7.2286 506 0.7593 0.7234 0.7593 0.8714
0.429 7.2571 508 0.7494 0.7324 0.7494 0.8657
0.429 7.2857 510 0.7764 0.7114 0.7764 0.8811
0.429 7.3143 512 0.8209 0.7089 0.8209 0.9061
0.429 7.3429 514 0.8410 0.7081 0.8410 0.9171
0.429 7.3714 516 0.7779 0.7362 0.7779 0.8820
0.429 7.4 518 0.7132 0.7297 0.7132 0.8445
0.429 7.4286 520 0.7198 0.7324 0.7198 0.8484
0.429 7.4571 522 0.7238 0.7324 0.7238 0.8508
0.429 7.4857 524 0.7071 0.7297 0.7071 0.8409
0.429 7.5143 526 0.7241 0.7211 0.7241 0.8509
0.429 7.5429 528 0.7437 0.7083 0.7437 0.8624
0.429 7.5714 530 0.7527 0.7234 0.7527 0.8676
0.429 7.6 532 0.7453 0.7324 0.7453 0.8633
0.429 7.6286 534 0.7437 0.7448 0.7437 0.8624
0.429 7.6571 536 0.7669 0.7286 0.7669 0.8757
0.429 7.6857 538 0.7607 0.7234 0.7607 0.8722
0.429 7.7143 540 0.7309 0.7310 0.7309 0.8549
0.429 7.7429 542 0.7675 0.7375 0.7675 0.8761
0.429 7.7714 544 0.8606 0.7073 0.8606 0.9277
0.429 7.8 546 0.8546 0.7073 0.8546 0.9245
0.429 7.8286 548 0.7805 0.6892 0.7805 0.8834
0.429 7.8571 550 0.7381 0.7361 0.7381 0.8591
0.429 7.8857 552 0.7567 0.7222 0.7567 0.8699
0.429 7.9143 554 0.7675 0.7222 0.7675 0.8761
0.429 7.9429 556 0.7427 0.7310 0.7427 0.8618
0.429 7.9714 558 0.7718 0.7362 0.7718 0.8785
0.429 8.0 560 0.9243 0.7209 0.9243 0.9614
0.429 8.0286 562 0.9933 0.7072 0.9933 0.9967
0.429 8.0571 564 0.9419 0.7263 0.9419 0.9705
0.429 8.0857 566 0.8331 0.7412 0.8331 0.9127
0.429 8.1143 568 0.7539 0.75 0.7539 0.8683
0.429 8.1429 570 0.7158 0.7361 0.7158 0.8461
0.429 8.1714 572 0.7277 0.6809 0.7277 0.8530
0.429 8.2 574 0.7973 0.6525 0.7973 0.8929
0.429 8.2286 576 0.8040 0.6389 0.8040 0.8967
0.429 8.2571 578 0.7681 0.6619 0.7681 0.8764
0.429 8.2857 580 0.7562 0.7 0.7562 0.8696

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k15_task1_organization

Finetuned
(4019)
this model