ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k6_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6883
  • Qwk: 0.6708
  • Mse: 0.6883
  • Rmse: 0.8297

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0625 2 5.1261 0.0052 5.1261 2.2641
No log 0.125 4 3.3783 0.0473 3.3783 1.8380
No log 0.1875 6 2.4724 -0.0827 2.4724 1.5724
No log 0.25 8 2.0075 0.0133 2.0075 1.4169
No log 0.3125 10 1.3888 0.1060 1.3888 1.1785
No log 0.375 12 1.2800 0.1988 1.2800 1.1314
No log 0.4375 14 1.2698 0.2361 1.2698 1.1268
No log 0.5 16 1.2698 0.2182 1.2698 1.1269
No log 0.5625 18 1.3019 0.1901 1.3019 1.1410
No log 0.625 20 1.3150 0.1292 1.3150 1.1468
No log 0.6875 22 1.2068 0.3036 1.2068 1.0985
No log 0.75 24 1.0922 0.3563 1.0922 1.0451
No log 0.8125 26 1.0133 0.3246 1.0133 1.0066
No log 0.875 28 1.0109 0.2735 1.0109 1.0054
No log 0.9375 30 0.9720 0.3025 0.9720 0.9859
No log 1.0 32 0.9681 0.3318 0.9681 0.9839
No log 1.0625 34 0.9612 0.3697 0.9612 0.9804
No log 1.125 36 1.0906 0.3613 1.0906 1.0443
No log 1.1875 38 1.2451 0.2425 1.2451 1.1159
No log 1.25 40 1.1455 0.3027 1.1455 1.0703
No log 1.3125 42 1.0657 0.4286 1.0657 1.0324
No log 1.375 44 0.9817 0.3808 0.9817 0.9908
No log 1.4375 46 1.1465 0.2827 1.1465 1.0707
No log 1.5 48 1.3753 0.2777 1.3753 1.1727
No log 1.5625 50 1.1831 0.3685 1.1831 1.0877
No log 1.625 52 0.9302 0.3817 0.9302 0.9645
No log 1.6875 54 0.8993 0.3659 0.8993 0.9483
No log 1.75 56 0.9200 0.4175 0.9200 0.9592
No log 1.8125 58 0.9792 0.4306 0.9792 0.9895
No log 1.875 60 1.0184 0.4291 1.0184 1.0092
No log 1.9375 62 0.9625 0.4737 0.9625 0.9811
No log 2.0 64 0.8820 0.4345 0.8820 0.9391
No log 2.0625 66 0.9511 0.4379 0.9511 0.9753
No log 2.125 68 0.9251 0.4363 0.9251 0.9618
No log 2.1875 70 0.8492 0.4455 0.8492 0.9215
No log 2.25 72 0.8255 0.4829 0.8255 0.9086
No log 2.3125 74 0.8080 0.4838 0.8080 0.8989
No log 2.375 76 0.8062 0.5081 0.8062 0.8979
No log 2.4375 78 0.7964 0.5257 0.7964 0.8924
No log 2.5 80 0.7837 0.5522 0.7837 0.8852
No log 2.5625 82 0.7192 0.6082 0.7192 0.8481
No log 2.625 84 0.7346 0.6011 0.7346 0.8571
No log 2.6875 86 0.8673 0.5690 0.8673 0.9313
No log 2.75 88 0.9683 0.5873 0.9683 0.9840
No log 2.8125 90 0.9606 0.5565 0.9606 0.9801
No log 2.875 92 0.8802 0.5446 0.8802 0.9382
No log 2.9375 94 0.7589 0.5756 0.7589 0.8712
No log 3.0 96 0.7587 0.5428 0.7587 0.8710
No log 3.0625 98 0.8623 0.5509 0.8623 0.9286
No log 3.125 100 0.7889 0.5355 0.7889 0.8882
No log 3.1875 102 0.6872 0.6288 0.6872 0.8290
No log 3.25 104 0.8077 0.6219 0.8077 0.8987
No log 3.3125 106 1.0683 0.4992 1.0683 1.0336
No log 3.375 108 1.0520 0.5258 1.0520 1.0257
No log 3.4375 110 0.8927 0.5711 0.8927 0.9448
No log 3.5 112 0.8743 0.5728 0.8743 0.9351
No log 3.5625 114 1.0678 0.5489 1.0678 1.0334
No log 3.625 116 1.0020 0.5780 1.0020 1.0010
No log 3.6875 118 0.7509 0.6235 0.7509 0.8666
No log 3.75 120 0.6827 0.6399 0.6827 0.8263
No log 3.8125 122 0.7622 0.5451 0.7622 0.8730
No log 3.875 124 0.9358 0.5757 0.9358 0.9674
No log 3.9375 126 1.1966 0.4708 1.1966 1.0939
No log 4.0 128 1.2544 0.4412 1.2544 1.1200
No log 4.0625 130 1.1471 0.4368 1.1471 1.0710
No log 4.125 132 0.9312 0.5131 0.9312 0.9650
No log 4.1875 134 0.7654 0.5679 0.7654 0.8749
No log 4.25 136 0.7771 0.5889 0.7771 0.8815
No log 4.3125 138 0.7894 0.5691 0.7894 0.8885
No log 4.375 140 0.7710 0.5777 0.7710 0.8780
No log 4.4375 142 0.7608 0.5922 0.7608 0.8722
No log 4.5 144 0.8262 0.5803 0.8262 0.9089
No log 4.5625 146 0.8017 0.5963 0.8017 0.8954
No log 4.625 148 0.7848 0.6083 0.7848 0.8859
No log 4.6875 150 0.7949 0.6076 0.7949 0.8916
No log 4.75 152 0.7656 0.5740 0.7656 0.8750
No log 4.8125 154 0.6976 0.6474 0.6976 0.8352
No log 4.875 156 0.6942 0.6738 0.6942 0.8332
No log 4.9375 158 0.7255 0.6050 0.7255 0.8518
No log 5.0 160 0.8481 0.5625 0.8481 0.9209
No log 5.0625 162 0.9532 0.5542 0.9532 0.9763
No log 5.125 164 1.0933 0.5487 1.0933 1.0456
No log 5.1875 166 1.0940 0.5539 1.0940 1.0460
No log 5.25 168 1.0120 0.5422 1.0120 1.0060
No log 5.3125 170 1.0247 0.5622 1.0247 1.0123
No log 5.375 172 0.9399 0.5852 0.9399 0.9695
No log 5.4375 174 0.8789 0.5986 0.8789 0.9375
No log 5.5 176 0.7798 0.6522 0.7798 0.8831
No log 5.5625 178 0.8426 0.6186 0.8426 0.9179
No log 5.625 180 1.0005 0.5496 1.0005 1.0002
No log 5.6875 182 1.0375 0.5816 1.0375 1.0186
No log 5.75 184 1.0672 0.5498 1.0672 1.0330
No log 5.8125 186 1.0143 0.5841 1.0143 1.0071
No log 5.875 188 0.8526 0.6098 0.8526 0.9234
No log 5.9375 190 0.7204 0.6302 0.7204 0.8488
No log 6.0 192 0.7337 0.6185 0.7337 0.8566
No log 6.0625 194 0.7156 0.6312 0.7156 0.8459
No log 6.125 196 0.7407 0.6430 0.7407 0.8607
No log 6.1875 198 0.8055 0.6483 0.8055 0.8975
No log 6.25 200 0.9380 0.5437 0.9380 0.9685
No log 6.3125 202 1.0160 0.5291 1.0160 1.0080
No log 6.375 204 0.9382 0.5755 0.9382 0.9686
No log 6.4375 206 0.9235 0.5755 0.9235 0.9610
No log 6.5 208 0.8988 0.5793 0.8988 0.9480
No log 6.5625 210 0.9443 0.5743 0.9443 0.9717
No log 6.625 212 1.0674 0.5317 1.0674 1.0331
No log 6.6875 214 1.0299 0.5149 1.0299 1.0148
No log 6.75 216 0.8243 0.5995 0.8243 0.9079
No log 6.8125 218 0.7592 0.6538 0.7592 0.8713
No log 6.875 220 0.7934 0.6250 0.7934 0.8908
No log 6.9375 222 0.8112 0.6305 0.8112 0.9007
No log 7.0 224 0.9611 0.5583 0.9611 0.9803
No log 7.0625 226 0.9686 0.5621 0.9686 0.9842
No log 7.125 228 0.9008 0.5679 0.9008 0.9491
No log 7.1875 230 0.8896 0.5717 0.8896 0.9432
No log 7.25 232 0.9191 0.5595 0.9191 0.9587
No log 7.3125 234 0.9048 0.5730 0.9048 0.9512
No log 7.375 236 0.7553 0.6577 0.7553 0.8691
No log 7.4375 238 0.6601 0.6922 0.6601 0.8125
No log 7.5 240 0.6913 0.6779 0.6913 0.8315
No log 7.5625 242 0.8194 0.5975 0.8194 0.9052
No log 7.625 244 0.8615 0.5795 0.8615 0.9282
No log 7.6875 246 0.9437 0.5792 0.9437 0.9714
No log 7.75 248 0.8911 0.6003 0.8911 0.9440
No log 7.8125 250 0.6904 0.6723 0.6904 0.8309
No log 7.875 252 0.6428 0.6722 0.6428 0.8017
No log 7.9375 254 0.6481 0.6992 0.6481 0.8051
No log 8.0 256 0.7109 0.6804 0.7109 0.8431
No log 8.0625 258 0.8203 0.6229 0.8203 0.9057
No log 8.125 260 0.9623 0.5473 0.9623 0.9810
No log 8.1875 262 0.9983 0.5381 0.9983 0.9992
No log 8.25 264 0.8711 0.5652 0.8711 0.9333
No log 8.3125 266 0.6941 0.6246 0.6941 0.8331
No log 8.375 268 0.6650 0.6300 0.6650 0.8155
No log 8.4375 270 0.6808 0.6162 0.6808 0.8251
No log 8.5 272 0.7295 0.6520 0.7295 0.8541
No log 8.5625 274 0.7746 0.6277 0.7746 0.8801
No log 8.625 276 0.7649 0.6215 0.7649 0.8746
No log 8.6875 278 0.7313 0.6189 0.7313 0.8551
No log 8.75 280 0.7037 0.6569 0.7037 0.8389
No log 8.8125 282 0.7302 0.6307 0.7302 0.8545
No log 8.875 284 0.8349 0.5821 0.8349 0.9137
No log 8.9375 286 0.9320 0.5501 0.9320 0.9654
No log 9.0 288 0.8914 0.5731 0.8914 0.9441
No log 9.0625 290 0.7563 0.6645 0.7563 0.8697
No log 9.125 292 0.6821 0.7094 0.6821 0.8259
No log 9.1875 294 0.6760 0.7212 0.6760 0.8222
No log 9.25 296 0.7616 0.6696 0.7616 0.8727
No log 9.3125 298 0.8055 0.6329 0.8055 0.8975
No log 9.375 300 0.8088 0.6588 0.8088 0.8993
No log 9.4375 302 0.8511 0.6116 0.8511 0.9225
No log 9.5 304 0.8202 0.6752 0.8202 0.9057
No log 9.5625 306 0.7366 0.6793 0.7366 0.8583
No log 9.625 308 0.7768 0.6628 0.7768 0.8813
No log 9.6875 310 0.7848 0.6628 0.7848 0.8859
No log 9.75 312 0.8122 0.6341 0.8122 0.9012
No log 9.8125 314 0.7802 0.6497 0.7802 0.8833
No log 9.875 316 0.7682 0.6271 0.7682 0.8765
No log 9.9375 318 0.8285 0.6205 0.8285 0.9102
No log 10.0 320 0.8372 0.6050 0.8372 0.9150
No log 10.0625 322 0.9278 0.6073 0.9278 0.9632
No log 10.125 324 0.9936 0.6164 0.9936 0.9968
No log 10.1875 326 0.9160 0.6203 0.9160 0.9571
No log 10.25 328 0.7619 0.6607 0.7619 0.8729
No log 10.3125 330 0.7354 0.6646 0.7354 0.8575
No log 10.375 332 0.7591 0.6337 0.7591 0.8712
No log 10.4375 334 0.8669 0.6094 0.8669 0.9311
No log 10.5 336 0.9007 0.5984 0.9007 0.9490
No log 10.5625 338 0.7929 0.6333 0.7929 0.8904
No log 10.625 340 0.6802 0.6558 0.6802 0.8248
No log 10.6875 342 0.6399 0.6874 0.6399 0.8000
No log 10.75 344 0.6572 0.6913 0.6572 0.8107
No log 10.8125 346 0.7548 0.6205 0.7548 0.8688
No log 10.875 348 1.0146 0.5858 1.0146 1.0073
No log 10.9375 350 1.1257 0.5666 1.1257 1.0610
No log 11.0 352 1.0293 0.5612 1.0293 1.0145
No log 11.0625 354 0.8382 0.5863 0.8382 0.9155
No log 11.125 356 0.6970 0.6302 0.6970 0.8348
No log 11.1875 358 0.6751 0.6475 0.6751 0.8217
No log 11.25 360 0.7414 0.6533 0.7414 0.8611
No log 11.3125 362 0.8649 0.6322 0.8649 0.9300
No log 11.375 364 0.8920 0.5932 0.8920 0.9445
No log 11.4375 366 0.7732 0.6824 0.7732 0.8793
No log 11.5 368 0.6922 0.7188 0.6922 0.8320
No log 11.5625 370 0.7003 0.6957 0.7003 0.8368
No log 11.625 372 0.8272 0.6346 0.8272 0.9095
No log 11.6875 374 0.8904 0.6030 0.8904 0.9436
No log 11.75 376 0.8791 0.5745 0.8791 0.9376
No log 11.8125 378 0.7967 0.6491 0.7967 0.8926
No log 11.875 380 0.7524 0.6474 0.7524 0.8674
No log 11.9375 382 0.7867 0.6353 0.7867 0.8870
No log 12.0 384 0.8256 0.6158 0.8256 0.9087
No log 12.0625 386 0.8854 0.5928 0.8854 0.9409
No log 12.125 388 0.8021 0.6465 0.8021 0.8956
No log 12.1875 390 0.6886 0.6813 0.6886 0.8298
No log 12.25 392 0.6516 0.6998 0.6516 0.8072
No log 12.3125 394 0.6314 0.7187 0.6314 0.7946
No log 12.375 396 0.6755 0.7536 0.6755 0.8219
No log 12.4375 398 0.7828 0.6792 0.7828 0.8848
No log 12.5 400 0.7699 0.6977 0.7699 0.8775
No log 12.5625 402 0.7518 0.7149 0.7518 0.8671
No log 12.625 404 0.8497 0.6396 0.8497 0.9218
No log 12.6875 406 0.8890 0.6063 0.8890 0.9429
No log 12.75 408 0.8567 0.6165 0.8567 0.9256
No log 12.8125 410 0.9406 0.6148 0.9406 0.9699
No log 12.875 412 1.0961 0.5986 1.0961 1.0470
No log 12.9375 414 1.1629 0.6051 1.1629 1.0784
No log 13.0 416 1.0713 0.5754 1.0713 1.0350
No log 13.0625 418 0.9495 0.5575 0.9495 0.9744
No log 13.125 420 0.8400 0.5969 0.8400 0.9165
No log 13.1875 422 0.8027 0.6562 0.8027 0.8960
No log 13.25 424 0.8267 0.6230 0.8267 0.9092
No log 13.3125 426 0.9283 0.5777 0.9283 0.9635
No log 13.375 428 0.9945 0.5433 0.9945 0.9973
No log 13.4375 430 0.9373 0.5708 0.9373 0.9681
No log 13.5 432 0.8191 0.6736 0.8191 0.9051
No log 13.5625 434 0.7386 0.6860 0.7386 0.8594
No log 13.625 436 0.7112 0.6809 0.7112 0.8433
No log 13.6875 438 0.6876 0.6859 0.6876 0.8292
No log 13.75 440 0.7879 0.6499 0.7879 0.8876
No log 13.8125 442 0.8713 0.6233 0.8713 0.9335
No log 13.875 444 0.8476 0.6193 0.8476 0.9206
No log 13.9375 446 0.7615 0.6458 0.7615 0.8727
No log 14.0 448 0.7056 0.6954 0.7056 0.8400
No log 14.0625 450 0.7159 0.7059 0.7159 0.8461
No log 14.125 452 0.7302 0.6816 0.7302 0.8545
No log 14.1875 454 0.7832 0.6493 0.7832 0.8850
No log 14.25 456 0.8135 0.6128 0.8135 0.9020
No log 14.3125 458 0.7450 0.6854 0.7450 0.8631
No log 14.375 460 0.6574 0.7144 0.6574 0.8108
No log 14.4375 462 0.6354 0.7261 0.6354 0.7971
No log 14.5 464 0.7207 0.7033 0.7207 0.8489
No log 14.5625 466 0.8950 0.6445 0.8950 0.9460
No log 14.625 468 0.9642 0.6338 0.9642 0.9819
No log 14.6875 470 0.9232 0.6176 0.9232 0.9608
No log 14.75 472 0.8134 0.6247 0.8134 0.9019
No log 14.8125 474 0.7128 0.6940 0.7128 0.8443
No log 14.875 476 0.6898 0.7036 0.6898 0.8305
No log 14.9375 478 0.7707 0.6586 0.7707 0.8779
No log 15.0 480 0.8940 0.5563 0.8940 0.9455
No log 15.0625 482 1.0267 0.5675 1.0267 1.0133
No log 15.125 484 0.9965 0.5675 0.9965 0.9983
No log 15.1875 486 0.8401 0.5777 0.8401 0.9166
No log 15.25 488 0.7878 0.6442 0.7878 0.8876
No log 15.3125 490 0.7879 0.6658 0.7879 0.8877
No log 15.375 492 0.7313 0.6735 0.7313 0.8551
No log 15.4375 494 0.6859 0.6837 0.6859 0.8282
No log 15.5 496 0.6097 0.7057 0.6097 0.7808
No log 15.5625 498 0.5791 0.6785 0.5791 0.7610
0.4225 15.625 500 0.5729 0.6785 0.5729 0.7569
0.4225 15.6875 502 0.5869 0.6825 0.5869 0.7661
0.4225 15.75 504 0.7313 0.6941 0.7313 0.8551
0.4225 15.8125 506 0.9840 0.6019 0.9840 0.9920
0.4225 15.875 508 1.0773 0.5869 1.0773 1.0379
0.4225 15.9375 510 0.9687 0.5933 0.9687 0.9842
0.4225 16.0 512 0.7527 0.6706 0.7527 0.8676
0.4225 16.0625 514 0.6044 0.6826 0.6044 0.7774
0.4225 16.125 516 0.5811 0.7072 0.5811 0.7623
0.4225 16.1875 518 0.5948 0.6870 0.5948 0.7712
0.4225 16.25 520 0.6488 0.6685 0.6488 0.8055
0.4225 16.3125 522 0.7241 0.6592 0.7241 0.8510
0.4225 16.375 524 0.7683 0.6387 0.7683 0.8765
0.4225 16.4375 526 0.7324 0.6647 0.7324 0.8558
0.4225 16.5 528 0.6883 0.6708 0.6883 0.8297

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run2_AugV5_k6_task1_organization

Finetuned
(4023)
this model