ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k11_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7886
  • Qwk: 0.7211
  • Mse: 0.7886
  • Rmse: 0.8880

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0392 2 6.8505 0.0176 6.8505 2.6174
No log 0.0784 4 4.3011 0.0779 4.3011 2.0739
No log 0.1176 6 3.3682 0.0111 3.3682 1.8353
No log 0.1569 8 2.6542 0.1096 2.6542 1.6292
No log 0.1961 10 2.4156 0.1449 2.4156 1.5542
No log 0.2353 12 2.1895 0.1159 2.1895 1.4797
No log 0.2745 14 2.5315 0.0268 2.5315 1.5911
No log 0.3137 16 3.0993 0.0732 3.0993 1.7605
No log 0.3529 18 3.0865 0.0982 3.0865 1.7569
No log 0.3922 20 2.3997 0.1299 2.3997 1.5491
No log 0.4314 22 1.5861 0.3252 1.5861 1.2594
No log 0.4706 24 1.4518 0.3158 1.4518 1.2049
No log 0.5098 26 1.5772 0.3000 1.5772 1.2559
No log 0.5490 28 2.0994 0.2254 2.0994 1.4489
No log 0.5882 30 2.4503 0.1006 2.4503 1.5654
No log 0.6275 32 2.2472 0.1497 2.2472 1.4991
No log 0.6667 34 1.9982 0.3000 1.9982 1.4136
No log 0.7059 36 1.7931 0.3134 1.7931 1.3390
No log 0.7451 38 1.6272 0.3511 1.6272 1.2756
No log 0.7843 40 1.6027 0.3623 1.6027 1.2660
No log 0.8235 42 1.4439 0.4444 1.4439 1.2016
No log 0.8627 44 1.3323 0.4412 1.3323 1.1543
No log 0.9020 46 1.3386 0.5 1.3386 1.1570
No log 0.9412 48 1.7163 0.4204 1.7163 1.3101
No log 0.9804 50 2.3787 0.3676 2.3787 1.5423
No log 1.0196 52 2.4032 0.3804 2.4032 1.5502
No log 1.0588 54 1.6003 0.4691 1.6003 1.2650
No log 1.0980 56 1.0033 0.5946 1.0033 1.0016
No log 1.1373 58 0.9993 0.5793 0.9993 0.9996
No log 1.1765 60 0.9582 0.6483 0.9582 0.9789
No log 1.2157 62 1.1575 0.6329 1.1575 1.0759
No log 1.2549 64 2.1387 0.3793 2.1387 1.4624
No log 1.2941 66 2.6569 0.2887 2.6569 1.6300
No log 1.3333 68 2.3868 0.3243 2.3868 1.5449
No log 1.3725 70 1.5039 0.4472 1.5039 1.2263
No log 1.4118 72 0.7942 0.7105 0.7942 0.8912
No log 1.4510 74 0.7190 0.72 0.7190 0.8479
No log 1.4902 76 0.7372 0.7421 0.7372 0.8586
No log 1.5294 78 1.0852 0.6118 1.0852 1.0417
No log 1.5686 80 1.6869 0.5474 1.6869 1.2988
No log 1.6078 82 2.0585 0.4523 2.0585 1.4348
No log 1.6471 84 2.0539 0.4677 2.0539 1.4331
No log 1.6863 86 1.7945 0.5291 1.7945 1.3396
No log 1.7255 88 1.3863 0.5652 1.3863 1.1774
No log 1.7647 90 1.2813 0.56 1.2813 1.1319
No log 1.8039 92 1.4327 0.5333 1.4327 1.1969
No log 1.8431 94 1.4279 0.5444 1.4279 1.1949
No log 1.8824 96 1.5187 0.5030 1.5187 1.2324
No log 1.9216 98 1.7029 0.5089 1.7029 1.3050
No log 1.9608 100 1.5320 0.5190 1.5320 1.2377
No log 2.0 102 1.2442 0.6027 1.2442 1.1154
No log 2.0392 104 1.1071 0.6234 1.1071 1.0522
No log 2.0784 106 1.2250 0.6433 1.2250 1.1068
No log 2.1176 108 1.0543 0.6706 1.0543 1.0268
No log 2.1569 110 0.9462 0.6746 0.9462 0.9727
No log 2.1961 112 0.9075 0.6941 0.9075 0.9526
No log 2.2353 114 0.8329 0.7126 0.8329 0.9126
No log 2.2745 116 0.8801 0.6936 0.8801 0.9381
No log 2.3137 118 1.1221 0.6854 1.1221 1.0593
No log 2.3529 120 1.6819 0.5 1.6819 1.2969
No log 2.3922 122 1.8699 0.4950 1.8699 1.3674
No log 2.4314 124 1.3375 0.5977 1.3375 1.1565
No log 2.4706 126 0.7872 0.7719 0.7872 0.8872
No log 2.5098 128 0.7054 0.7564 0.7054 0.8399
No log 2.5490 130 0.6942 0.7578 0.6942 0.8332
No log 2.5882 132 0.7613 0.7738 0.7613 0.8725
No log 2.6275 134 0.7806 0.7692 0.7806 0.8835
No log 2.6667 136 1.1083 0.6286 1.1083 1.0527
No log 2.7059 138 1.8503 0.5146 1.8503 1.3602
No log 2.7451 140 1.9437 0.5352 1.9437 1.3942
No log 2.7843 142 1.3237 0.6263 1.3237 1.1505
No log 2.8235 144 0.8858 0.7514 0.8858 0.9412
No log 2.8627 146 0.9329 0.6748 0.9329 0.9659
No log 2.9020 148 1.0349 0.6242 1.0349 1.0173
No log 2.9412 150 0.9511 0.6375 0.9511 0.9752
No log 2.9804 152 0.8156 0.7841 0.8156 0.9031
No log 3.0196 154 0.9294 0.7072 0.9294 0.9640
No log 3.0588 156 0.8540 0.7251 0.8540 0.9241
No log 3.0980 158 0.6889 0.7532 0.6889 0.8300
No log 3.1373 160 0.6722 0.7260 0.6722 0.8199
No log 3.1765 162 0.6656 0.7123 0.6656 0.8159
No log 3.2157 164 0.6714 0.7355 0.6714 0.8194
No log 3.2549 166 0.8956 0.7176 0.8956 0.9464
No log 3.2941 168 1.3609 0.6131 1.3609 1.1666
No log 3.3333 170 1.4408 0.5911 1.4408 1.2003
No log 3.3725 172 1.0825 0.6904 1.0825 1.0404
No log 3.4118 174 0.7427 0.7791 0.7427 0.8618
No log 3.4510 176 0.6915 0.7950 0.6915 0.8315
No log 3.4902 178 0.6482 0.8 0.6482 0.8051
No log 3.5294 180 0.6348 0.7882 0.6348 0.7967
No log 3.5686 182 0.8048 0.75 0.8048 0.8971
No log 3.6078 184 0.8152 0.75 0.8152 0.9029
No log 3.6471 186 0.6870 0.7574 0.6870 0.8288
No log 3.6863 188 0.6914 0.7722 0.6914 0.8315
No log 3.7255 190 0.7882 0.72 0.7882 0.8878
No log 3.7647 192 0.8850 0.6434 0.8850 0.9407
No log 3.8039 194 0.7498 0.7114 0.7498 0.8659
No log 3.8431 196 0.6492 0.7632 0.6492 0.8058
No log 3.8824 198 0.7246 0.7778 0.7246 0.8513
No log 3.9216 200 0.9413 0.6867 0.9413 0.9702
No log 3.9608 202 1.1497 0.6286 1.1497 1.0722
No log 4.0 204 1.0987 0.6667 1.0987 1.0482
No log 4.0392 206 0.9353 0.7219 0.9353 0.9671
No log 4.0784 208 0.8253 0.7375 0.8253 0.9085
No log 4.1176 210 0.7673 0.7683 0.7673 0.8759
No log 4.1569 212 0.8183 0.7368 0.8183 0.9046
No log 4.1961 214 0.8015 0.7674 0.8015 0.8952
No log 4.2353 216 0.7898 0.7746 0.7898 0.8887
No log 4.2745 218 0.6709 0.7976 0.6709 0.8191
No log 4.3137 220 0.6432 0.7632 0.6432 0.8020
No log 4.3529 222 0.6845 0.7162 0.6845 0.8273
No log 4.3922 224 0.8037 0.7285 0.8037 0.8965
No log 4.4314 226 0.9987 0.6832 0.9987 0.9993
No log 4.4706 228 1.0422 0.6707 1.0422 1.0209
No log 4.5098 230 0.9120 0.6918 0.9120 0.9550
No log 4.5490 232 0.8019 0.7389 0.8019 0.8955
No log 4.5882 234 0.6923 0.7595 0.6923 0.8321
No log 4.6275 236 0.6862 0.7758 0.6862 0.8284
No log 4.6667 238 0.7504 0.7262 0.7504 0.8663
No log 4.7059 240 0.8542 0.7399 0.8542 0.9242
No log 4.7451 242 0.8360 0.7399 0.8360 0.9143
No log 4.7843 244 0.8731 0.7251 0.8731 0.9344
No log 4.8235 246 0.8359 0.7381 0.8359 0.9142
No log 4.8627 248 0.7635 0.7425 0.7635 0.8738
No log 4.9020 250 0.6914 0.7273 0.6914 0.8315
No log 4.9412 252 0.7052 0.7517 0.7052 0.8398
No log 4.9804 254 0.7053 0.7286 0.7053 0.8398
No log 5.0196 256 0.7015 0.7429 0.7015 0.8375
No log 5.0588 258 0.7266 0.7397 0.7266 0.8524
No log 5.0980 260 0.7711 0.7285 0.7711 0.8781
No log 5.1373 262 0.8241 0.6792 0.8241 0.9078
No log 5.1765 264 1.0920 0.6353 1.0920 1.0450
No log 5.2157 266 1.1956 0.6339 1.1956 1.0934
No log 5.2549 268 0.8750 0.7528 0.8750 0.9354
No log 5.2941 270 0.6074 0.7882 0.6074 0.7794
No log 5.3333 272 0.6009 0.7515 0.6009 0.7752
No log 5.3725 274 0.7118 0.7861 0.7118 0.8437
No log 5.4118 276 0.9569 0.7174 0.9569 0.9782
No log 5.4510 278 0.9419 0.7303 0.9419 0.9705
No log 5.4902 280 0.7203 0.7574 0.7203 0.8487
No log 5.5294 282 0.6197 0.7613 0.6197 0.7872
No log 5.5686 284 0.6561 0.7273 0.6561 0.8100
No log 5.6078 286 0.6749 0.7143 0.6749 0.8215
No log 5.6471 288 0.6662 0.7534 0.6662 0.8162
No log 5.6863 290 0.7738 0.7389 0.7738 0.8796
No log 5.7255 292 0.8468 0.7219 0.8468 0.9202
No log 5.7647 294 0.7796 0.7590 0.7796 0.8830
No log 5.8039 296 0.7328 0.7485 0.7328 0.8561
No log 5.8431 298 0.6590 0.7950 0.6590 0.8118
No log 5.8824 300 0.6529 0.7517 0.6529 0.8080
No log 5.9216 302 0.6885 0.7347 0.6885 0.8297
No log 5.9608 304 0.8061 0.7114 0.8061 0.8978
No log 6.0 306 0.9647 0.6174 0.9647 0.9822
No log 6.0392 308 0.9415 0.6623 0.9415 0.9703
No log 6.0784 310 0.8572 0.7020 0.8572 0.9258
No log 6.1176 312 0.8085 0.7285 0.8085 0.8992
No log 6.1569 314 0.7416 0.7237 0.7416 0.8612
No log 6.1961 316 0.6584 0.7643 0.6584 0.8114
No log 6.2353 318 0.7030 0.7545 0.7030 0.8384
No log 6.2745 320 0.9237 0.7380 0.9237 0.9611
No log 6.3137 322 1.1512 0.6839 1.1512 1.0730
No log 6.3529 324 1.1356 0.6875 1.1356 1.0657
No log 6.3922 326 0.8812 0.7416 0.8812 0.9387
No log 6.4314 328 0.6776 0.7730 0.6776 0.8232
No log 6.4706 330 0.6665 0.7162 0.6665 0.8164
No log 6.5098 332 0.6972 0.6849 0.6972 0.8350
No log 6.5490 334 0.8539 0.6939 0.8539 0.9241
No log 6.5882 336 1.0415 0.6267 1.0415 1.0205
No log 6.6275 338 1.2050 0.6242 1.2050 1.0977
No log 6.6667 340 1.1453 0.6543 1.1453 1.0702
No log 6.7059 342 0.9763 0.7089 0.9763 0.9881
No log 6.7451 344 0.8627 0.6968 0.8627 0.9288
No log 6.7843 346 0.8823 0.7073 0.8823 0.9393
No log 6.8235 348 0.8880 0.7186 0.8880 0.9423
No log 6.8627 350 0.7347 0.7484 0.7347 0.8572
No log 6.9020 352 0.6609 0.7682 0.6609 0.8130
No log 6.9412 354 0.6656 0.7034 0.6656 0.8159
No log 6.9804 356 0.7224 0.7042 0.7224 0.8499
No log 7.0196 358 0.7890 0.6715 0.7890 0.8882
No log 7.0588 360 0.8912 0.5970 0.8912 0.9440
No log 7.0980 362 0.9029 0.5821 0.9029 0.9502
No log 7.1373 364 0.8241 0.6667 0.8241 0.9078
No log 7.1765 366 0.7632 0.7237 0.7632 0.8736
No log 7.2157 368 0.7427 0.7532 0.7427 0.8618
No log 7.2549 370 0.7820 0.7439 0.7820 0.8843
No log 7.2941 372 0.7741 0.7558 0.7741 0.8798
No log 7.3333 374 0.8675 0.7473 0.8675 0.9314
No log 7.3725 376 0.8037 0.7528 0.8037 0.8965
No log 7.4118 378 0.7139 0.7640 0.7139 0.8449
No log 7.4510 380 0.7708 0.7458 0.7708 0.8780
No log 7.4902 382 1.0177 0.7293 1.0177 1.0088
No log 7.5294 384 1.1735 0.6702 1.1735 1.0833
No log 7.5686 386 1.0658 0.6932 1.0658 1.0324
No log 7.6078 388 0.8902 0.7436 0.8902 0.9435
No log 7.6471 390 0.8015 0.7123 0.8015 0.8953
No log 7.6863 392 0.8717 0.7042 0.8717 0.9336
No log 7.7255 394 1.0078 0.6286 1.0078 1.0039
No log 7.7647 396 1.0058 0.6434 1.0058 1.0029
No log 7.8039 398 0.9270 0.6667 0.9270 0.9628
No log 7.8431 400 0.7989 0.7083 0.7989 0.8938
No log 7.8824 402 0.7120 0.7483 0.7120 0.8438
No log 7.9216 404 0.6869 0.76 0.6869 0.8288
No log 7.9608 406 0.7053 0.7632 0.7053 0.8398
No log 8.0 408 0.7042 0.7632 0.7042 0.8392
No log 8.0392 410 0.7800 0.7211 0.7800 0.8832
No log 8.0784 412 0.8947 0.7320 0.8947 0.9459
No log 8.1176 414 0.9214 0.7105 0.9214 0.9599
No log 8.1569 416 0.8591 0.7308 0.8591 0.9269
No log 8.1961 418 0.8640 0.7308 0.8640 0.9295
No log 8.2353 420 0.8326 0.7273 0.8326 0.9125
No log 8.2745 422 0.8094 0.7262 0.8094 0.8997
No log 8.3137 424 0.7356 0.7529 0.7356 0.8577
No log 8.3529 426 0.6179 0.7719 0.6179 0.7861
No log 8.3922 428 0.5889 0.7901 0.5889 0.7674
No log 8.4314 430 0.6244 0.7901 0.6244 0.7902
No log 8.4706 432 0.6910 0.7516 0.6910 0.8313
No log 8.5098 434 0.7653 0.7453 0.7653 0.8748
No log 8.5490 436 0.7981 0.7595 0.7981 0.8934
No log 8.5882 438 0.8015 0.7436 0.8015 0.8953
No log 8.6275 440 0.7505 0.7517 0.7505 0.8663
No log 8.6667 442 0.7291 0.7448 0.7291 0.8539
No log 8.7059 444 0.7257 0.7448 0.7257 0.8519
No log 8.7451 446 0.7125 0.7361 0.7125 0.8441
No log 8.7843 448 0.6359 0.7483 0.6359 0.7974
No log 8.8235 450 0.6063 0.7651 0.6063 0.7787
No log 8.8627 452 0.6685 0.7451 0.6685 0.8176
No log 8.9020 454 0.6922 0.7368 0.6922 0.8320
No log 8.9412 456 0.6145 0.7821 0.6145 0.7839
No log 8.9804 458 0.5476 0.7771 0.5476 0.7400
No log 9.0196 460 0.5437 0.7925 0.5437 0.7374
No log 9.0588 462 0.5894 0.7898 0.5894 0.7677
No log 9.0980 464 0.7939 0.7097 0.7939 0.8910
No log 9.1373 466 0.9894 0.6625 0.9894 0.9947
No log 9.1765 468 1.0178 0.6625 1.0178 1.0089
No log 9.2157 470 0.8833 0.7097 0.8833 0.9398
No log 9.2549 472 0.7745 0.7097 0.7745 0.8801
No log 9.2941 474 0.7652 0.7143 0.7652 0.8747
No log 9.3333 476 0.8011 0.7143 0.8011 0.8950
No log 9.3725 478 0.9183 0.6875 0.9183 0.9583
No log 9.4118 480 0.9878 0.6909 0.9878 0.9939
No log 9.4510 482 0.9573 0.6867 0.9573 0.9784
No log 9.4902 484 0.8271 0.6871 0.8271 0.9095
No log 9.5294 486 0.7327 0.7607 0.7327 0.8560
No log 9.5686 488 0.7003 0.7758 0.7003 0.8369
No log 9.6078 490 0.6230 0.7904 0.6230 0.7893
No log 9.6471 492 0.6114 0.7904 0.6114 0.7819
No log 9.6863 494 0.6476 0.7831 0.6476 0.8047
No log 9.7255 496 0.6847 0.7831 0.6847 0.8275
No log 9.7647 498 0.6637 0.7758 0.6637 0.8147
0.4425 9.8039 500 0.7027 0.7692 0.7027 0.8383
0.4425 9.8431 502 0.6583 0.7815 0.6583 0.8114
0.4425 9.8824 504 0.6021 0.7733 0.6021 0.7760
0.4425 9.9216 506 0.5816 0.7925 0.5816 0.7626
0.4425 9.9608 508 0.5580 0.7925 0.5580 0.7470
0.4425 10.0 510 0.5414 0.7875 0.5414 0.7358
0.4425 10.0392 512 0.5330 0.8125 0.5330 0.7301
0.4425 10.0784 514 0.5287 0.7898 0.5287 0.7271
0.4425 10.1176 516 0.5389 0.8075 0.5389 0.7341
0.4425 10.1569 518 0.5769 0.7875 0.5769 0.7595
0.4425 10.1961 520 0.6831 0.7831 0.6831 0.8265
0.4425 10.2353 522 0.8229 0.7284 0.8229 0.9071
0.4425 10.2745 524 0.8719 0.7117 0.8719 0.9338
0.4425 10.3137 526 0.8079 0.7485 0.8079 0.8988
0.4425 10.3529 528 0.6472 0.7831 0.6472 0.8045
0.4425 10.3922 530 0.5762 0.7929 0.5762 0.7591
0.4425 10.4314 532 0.5873 0.7831 0.5873 0.7663
0.4425 10.4706 534 0.6667 0.7831 0.6667 0.8165
0.4425 10.5098 536 0.8072 0.7329 0.8072 0.8984
0.4425 10.5490 538 1.0119 0.6667 1.0119 1.0059
0.4425 10.5882 540 1.0815 0.6471 1.0815 1.0399
0.4425 10.6275 542 0.9633 0.6709 0.9633 0.9815
0.4425 10.6667 544 0.7886 0.7211 0.7886 0.8880

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k11_task1_organization

Finetuned
(4019)
this model