ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7895
  • Qwk: 0.5236
  • Mse: 0.7895
  • Rmse: 0.8885

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0952 2 4.0904 -0.0331 4.0904 2.0225
No log 0.1905 4 2.6377 -0.0270 2.6377 1.6241
No log 0.2857 6 1.9230 -0.0905 1.9230 1.3867
No log 0.3810 8 1.4281 0.0389 1.4281 1.1950
No log 0.4762 10 1.2102 0.1142 1.2102 1.1001
No log 0.5714 12 1.2256 0.0380 1.2256 1.1071
No log 0.6667 14 1.1346 0.0318 1.1346 1.0652
No log 0.7619 16 1.0032 0.2935 1.0032 1.0016
No log 0.8571 18 0.9721 0.3709 0.9721 0.9859
No log 0.9524 20 0.9949 0.3160 0.9949 0.9975
No log 1.0476 22 1.1297 0.3299 1.1297 1.0629
No log 1.1429 24 1.0481 0.3270 1.0481 1.0238
No log 1.2381 26 1.0828 0.2455 1.0828 1.0406
No log 1.3333 28 1.1826 0.1671 1.1826 1.0875
No log 1.4286 30 1.2654 0.0914 1.2654 1.1249
No log 1.5238 32 1.2362 0.1115 1.2362 1.1119
No log 1.6190 34 1.1685 0.1568 1.1685 1.0810
No log 1.7143 36 1.1682 0.2777 1.1682 1.0808
No log 1.8095 38 1.1485 0.2100 1.1485 1.0717
No log 1.9048 40 1.1899 0.1958 1.1899 1.0908
No log 2.0 42 1.1535 0.2375 1.1535 1.0740
No log 2.0952 44 1.0608 0.2871 1.0608 1.0300
No log 2.1905 46 1.0238 0.3258 1.0238 1.0118
No log 2.2857 48 1.0370 0.3563 1.0370 1.0183
No log 2.3810 50 0.9900 0.3284 0.9900 0.9950
No log 2.4762 52 1.0261 0.3559 1.0261 1.0130
No log 2.5714 54 1.2236 0.3863 1.2236 1.1062
No log 2.6667 56 1.3208 0.3108 1.3208 1.1493
No log 2.7619 58 1.1483 0.2925 1.1483 1.0716
No log 2.8571 60 0.9858 0.2424 0.9858 0.9929
No log 2.9524 62 0.9899 0.2541 0.9899 0.9949
No log 3.0476 64 1.0242 0.2416 1.0242 1.0120
No log 3.1429 66 1.0271 0.2391 1.0271 1.0134
No log 3.2381 68 0.9765 0.3631 0.9765 0.9882
No log 3.3333 70 0.9241 0.3838 0.9241 0.9613
No log 3.4286 72 0.8842 0.3858 0.8842 0.9403
No log 3.5238 74 0.8915 0.4478 0.8915 0.9442
No log 3.6190 76 0.9596 0.3844 0.9596 0.9796
No log 3.7143 78 1.0432 0.3734 1.0432 1.0214
No log 3.8095 80 0.9403 0.4590 0.9403 0.9697
No log 3.9048 82 0.8867 0.4395 0.8867 0.9416
No log 4.0 84 0.8545 0.4996 0.8545 0.9244
No log 4.0952 86 0.8631 0.4577 0.8631 0.9290
No log 4.1905 88 0.8756 0.4563 0.8756 0.9357
No log 4.2857 90 0.8722 0.5110 0.8722 0.9339
No log 4.3810 92 0.9064 0.5188 0.9064 0.9520
No log 4.4762 94 0.9267 0.4420 0.9267 0.9626
No log 4.5714 96 0.8701 0.4305 0.8701 0.9328
No log 4.6667 98 0.9087 0.4708 0.9087 0.9533
No log 4.7619 100 0.7791 0.5404 0.7791 0.8827
No log 4.8571 102 0.7619 0.6070 0.7619 0.8729
No log 4.9524 104 0.8593 0.5420 0.8593 0.9270
No log 5.0476 106 0.7841 0.6682 0.7841 0.8855
No log 5.1429 108 0.8663 0.4935 0.8663 0.9307
No log 5.2381 110 1.1425 0.4208 1.1425 1.0689
No log 5.3333 112 1.1274 0.4788 1.1274 1.0618
No log 5.4286 114 0.8349 0.5247 0.8349 0.9137
No log 5.5238 116 0.7903 0.6143 0.7903 0.8890
No log 5.6190 118 0.7888 0.6053 0.7888 0.8881
No log 5.7143 120 0.9111 0.4751 0.9111 0.9545
No log 5.8095 122 1.0298 0.4432 1.0298 1.0148
No log 5.9048 124 0.8523 0.5257 0.8523 0.9232
No log 6.0 126 0.7961 0.5178 0.7961 0.8922
No log 6.0952 128 0.7978 0.5481 0.7978 0.8932
No log 6.1905 130 0.8704 0.4747 0.8704 0.9329
No log 6.2857 132 0.9620 0.4219 0.9620 0.9808
No log 6.3810 134 0.9022 0.4552 0.9022 0.9498
No log 6.4762 136 0.8664 0.4855 0.8664 0.9308
No log 6.5714 138 0.8585 0.5395 0.8585 0.9265
No log 6.6667 140 0.8306 0.5171 0.8306 0.9114
No log 6.7619 142 0.8162 0.4949 0.8162 0.9035
No log 6.8571 144 0.8067 0.4819 0.8067 0.8982
No log 6.9524 146 0.7978 0.5419 0.7978 0.8932
No log 7.0476 148 0.8275 0.4556 0.8275 0.9097
No log 7.1429 150 0.7904 0.6002 0.7904 0.8891
No log 7.2381 152 0.8119 0.6223 0.8119 0.9010
No log 7.3333 154 0.8853 0.5293 0.8853 0.9409
No log 7.4286 156 0.8814 0.5102 0.8814 0.9388
No log 7.5238 158 0.7744 0.6100 0.7744 0.8800
No log 7.6190 160 0.7379 0.5884 0.7379 0.8590
No log 7.7143 162 0.7502 0.5275 0.7502 0.8661
No log 7.8095 164 0.7676 0.5552 0.7676 0.8762
No log 7.9048 166 0.8413 0.5156 0.8413 0.9172
No log 8.0 168 0.8036 0.5737 0.8036 0.8965
No log 8.0952 170 0.7777 0.6422 0.7777 0.8819
No log 8.1905 172 0.7728 0.6606 0.7728 0.8791
No log 8.2857 174 0.7625 0.6342 0.7625 0.8732
No log 8.3810 176 0.7548 0.6390 0.7548 0.8688
No log 8.4762 178 0.7276 0.6511 0.7276 0.8530
No log 8.5714 180 0.7227 0.6484 0.7227 0.8501
No log 8.6667 182 0.7425 0.6148 0.7425 0.8617
No log 8.7619 184 0.7162 0.6319 0.7162 0.8463
No log 8.8571 186 0.7303 0.6224 0.7303 0.8546
No log 8.9524 188 0.7624 0.5586 0.7624 0.8732
No log 9.0476 190 0.7751 0.5614 0.7751 0.8804
No log 9.1429 192 0.7847 0.5495 0.7847 0.8858
No log 9.2381 194 0.7873 0.5495 0.7873 0.8873
No log 9.3333 196 0.7823 0.5328 0.7823 0.8845
No log 9.4286 198 0.8135 0.4964 0.8135 0.9020
No log 9.5238 200 0.7722 0.5316 0.7722 0.8787
No log 9.6190 202 0.7312 0.5763 0.7312 0.8551
No log 9.7143 204 0.7269 0.5929 0.7269 0.8526
No log 9.8095 206 0.8522 0.5109 0.8522 0.9231
No log 9.9048 208 0.8825 0.5098 0.8825 0.9394
No log 10.0 210 0.8738 0.5283 0.8738 0.9348
No log 10.0952 212 0.9797 0.5166 0.9797 0.9898
No log 10.1905 214 0.9143 0.4387 0.9143 0.9562
No log 10.2857 216 0.7763 0.5419 0.7763 0.8811
No log 10.3810 218 0.7571 0.5015 0.7571 0.8701
No log 10.4762 220 0.7591 0.4572 0.7591 0.8713
No log 10.5714 222 0.7896 0.5528 0.7896 0.8886
No log 10.6667 224 0.8495 0.4929 0.8495 0.9217
No log 10.7619 226 0.7950 0.5933 0.7950 0.8916
No log 10.8571 228 0.7666 0.5252 0.7666 0.8755
No log 10.9524 230 0.8097 0.5524 0.8097 0.8998
No log 11.0476 232 0.7949 0.5524 0.7949 0.8916
No log 11.1429 234 0.7484 0.5618 0.7484 0.8651
No log 11.2381 236 0.8442 0.5614 0.8442 0.9188
No log 11.3333 238 0.9247 0.4902 0.9247 0.9616
No log 11.4286 240 0.8255 0.5706 0.8255 0.9086
No log 11.5238 242 0.7320 0.6235 0.7320 0.8556
No log 11.6190 244 0.8093 0.4932 0.8093 0.8996
No log 11.7143 246 0.8438 0.4939 0.8438 0.9186
No log 11.8095 248 0.7827 0.4798 0.7827 0.8847
No log 11.9048 250 0.7566 0.5682 0.7566 0.8698
No log 12.0 252 0.7636 0.5124 0.7636 0.8738
No log 12.0952 254 0.7596 0.5569 0.7596 0.8715
No log 12.1905 256 0.7551 0.5569 0.7551 0.8690
No log 12.2857 258 0.7524 0.5250 0.7524 0.8674
No log 12.3810 260 0.7618 0.5236 0.7618 0.8728
No log 12.4762 262 0.8255 0.5385 0.8255 0.9086
No log 12.5714 264 0.8553 0.4478 0.8553 0.9248
No log 12.6667 266 0.8070 0.5392 0.8070 0.8983
No log 12.7619 268 0.7712 0.5439 0.7712 0.8782
No log 12.8571 270 0.7664 0.5763 0.7664 0.8754
No log 12.9524 272 0.7752 0.4989 0.7752 0.8804
No log 13.0476 274 0.8495 0.5131 0.8495 0.9217
No log 13.1429 276 0.8363 0.5365 0.8363 0.9145
No log 13.2381 278 0.7911 0.5069 0.7911 0.8894
No log 13.3333 280 0.7777 0.4691 0.7777 0.8819
No log 13.4286 282 0.7776 0.4691 0.7776 0.8818
No log 13.5238 284 0.7715 0.4133 0.7715 0.8783
No log 13.6190 286 0.7744 0.3941 0.7744 0.8800
No log 13.7143 288 0.7721 0.4345 0.7721 0.8787
No log 13.8095 290 0.7598 0.4145 0.7598 0.8717
No log 13.9048 292 0.7757 0.4960 0.7757 0.8808
No log 14.0 294 0.8065 0.4601 0.8065 0.8980
No log 14.0952 296 0.7967 0.5156 0.7967 0.8926
No log 14.1905 298 0.7606 0.5505 0.7606 0.8721
No log 14.2857 300 0.7346 0.6365 0.7346 0.8571
No log 14.3810 302 0.7262 0.5874 0.7262 0.8522
No log 14.4762 304 0.7201 0.5863 0.7201 0.8486
No log 14.5714 306 0.7336 0.5221 0.7336 0.8565
No log 14.6667 308 0.7315 0.5017 0.7315 0.8553
No log 14.7619 310 0.7232 0.5261 0.7232 0.8504
No log 14.8571 312 0.7254 0.5863 0.7254 0.8517
No log 14.9524 314 0.7436 0.6129 0.7436 0.8623
No log 15.0476 316 0.7694 0.6142 0.7694 0.8772
No log 15.1429 318 0.7802 0.6116 0.7802 0.8833
No log 15.2381 320 0.7927 0.6026 0.7927 0.8903
No log 15.3333 322 0.8315 0.5736 0.8315 0.9119
No log 15.4286 324 0.8374 0.5652 0.8374 0.9151
No log 15.5238 326 0.8447 0.5458 0.8447 0.9191
No log 15.6190 328 0.7986 0.5069 0.7986 0.8936
No log 15.7143 330 0.7657 0.5250 0.7657 0.8751
No log 15.8095 332 0.7608 0.5614 0.7608 0.8723
No log 15.9048 334 0.7623 0.5610 0.7623 0.8731
No log 16.0 336 0.7589 0.5592 0.7589 0.8711
No log 16.0952 338 0.8020 0.5885 0.8020 0.8955
No log 16.1905 340 0.7938 0.6064 0.7938 0.8909
No log 16.2857 342 0.7319 0.5603 0.7319 0.8555
No log 16.3810 344 0.7045 0.6423 0.7045 0.8393
No log 16.4762 346 0.7177 0.5974 0.7177 0.8472
No log 16.5714 348 0.7178 0.6215 0.7178 0.8472
No log 16.6667 350 0.7760 0.6217 0.7760 0.8809
No log 16.7619 352 0.7776 0.6217 0.7776 0.8818
No log 16.8571 354 0.7449 0.5869 0.7449 0.8631
No log 16.9524 356 0.7265 0.5711 0.7265 0.8524
No log 17.0476 358 0.7184 0.5840 0.7184 0.8476
No log 17.1429 360 0.7327 0.5691 0.7327 0.8560
No log 17.2381 362 0.7819 0.6057 0.7819 0.8843
No log 17.3333 364 0.8554 0.5222 0.8554 0.9249
No log 17.4286 366 0.8150 0.5433 0.8150 0.9028
No log 17.5238 368 0.7332 0.5905 0.7332 0.8563
No log 17.6190 370 0.7017 0.6398 0.7017 0.8377
No log 17.7143 372 0.6950 0.6398 0.6950 0.8337
No log 17.8095 374 0.7215 0.6254 0.7215 0.8494
No log 17.9048 376 0.7536 0.6318 0.7536 0.8681
No log 18.0 378 0.7953 0.6545 0.7953 0.8918
No log 18.0952 380 0.7992 0.6642 0.7992 0.8940
No log 18.1905 382 0.7500 0.6427 0.7500 0.8661
No log 18.2857 384 0.7029 0.6350 0.7029 0.8384
No log 18.3810 386 0.7011 0.6022 0.7011 0.8373
No log 18.4762 388 0.6922 0.5843 0.6922 0.8320
No log 18.5714 390 0.7059 0.5866 0.7059 0.8402
No log 18.6667 392 0.7429 0.5993 0.7429 0.8619
No log 18.7619 394 0.7235 0.6451 0.7235 0.8506
No log 18.8571 396 0.6978 0.6280 0.6978 0.8354
No log 18.9524 398 0.6864 0.6414 0.6864 0.8285
No log 19.0476 400 0.7139 0.6503 0.7139 0.8449
No log 19.1429 402 0.7294 0.6503 0.7294 0.8540
No log 19.2381 404 0.7458 0.6225 0.7458 0.8636
No log 19.3333 406 0.7401 0.5963 0.7401 0.8603
No log 19.4286 408 0.7406 0.6169 0.7406 0.8606
No log 19.5238 410 0.7405 0.5818 0.7405 0.8605
No log 19.6190 412 0.7643 0.6012 0.7643 0.8742
No log 19.7143 414 0.7864 0.5559 0.7864 0.8868
No log 19.8095 416 0.8240 0.5717 0.8240 0.9078
No log 19.9048 418 0.7989 0.5532 0.7989 0.8938
No log 20.0 420 0.7367 0.6151 0.7367 0.8583
No log 20.0952 422 0.7047 0.5827 0.7047 0.8394
No log 20.1905 424 0.7006 0.5827 0.7006 0.8370
No log 20.2857 426 0.7292 0.5691 0.7292 0.8539
No log 20.3810 428 0.8449 0.5353 0.8449 0.9192
No log 20.4762 430 0.9063 0.5006 0.9063 0.9520
No log 20.5714 432 0.8432 0.4823 0.8432 0.9183
No log 20.6667 434 0.7753 0.5002 0.7753 0.8805
No log 20.7619 436 0.7969 0.5542 0.7969 0.8927
No log 20.8571 438 0.8154 0.4870 0.8154 0.9030
No log 20.9524 440 0.7966 0.5176 0.7966 0.8925
No log 21.0476 442 0.7846 0.5142 0.7846 0.8858
No log 21.1429 444 0.7963 0.4876 0.7963 0.8923
No log 21.2381 446 0.7930 0.5279 0.7930 0.8905
No log 21.3333 448 0.7615 0.5700 0.7615 0.8726
No log 21.4286 450 0.7424 0.5921 0.7424 0.8616
No log 21.5238 452 0.7400 0.6148 0.7400 0.8603
No log 21.6190 454 0.7411 0.6148 0.7411 0.8609
No log 21.7143 456 0.7460 0.6120 0.7460 0.8637
No log 21.8095 458 0.7519 0.6235 0.7519 0.8671
No log 21.9048 460 0.7661 0.5700 0.7661 0.8753
No log 22.0 462 0.7839 0.5690 0.7839 0.8854
No log 22.0952 464 0.7931 0.5390 0.7931 0.8906
No log 22.1905 466 0.8008 0.5279 0.8008 0.8949
No log 22.2857 468 0.7877 0.5517 0.7877 0.8875
No log 22.3810 470 0.7766 0.4799 0.7766 0.8812
No log 22.4762 472 0.7756 0.5166 0.7756 0.8807
No log 22.5714 474 0.7737 0.5044 0.7737 0.8796
No log 22.6667 476 0.7815 0.5317 0.7815 0.8840
No log 22.7619 478 0.7953 0.5279 0.7953 0.8918
No log 22.8571 480 0.8033 0.5746 0.8033 0.8963
No log 22.9524 482 0.8224 0.5850 0.8224 0.9068
No log 23.0476 484 0.8205 0.6358 0.8205 0.9058
No log 23.1429 486 0.7623 0.6041 0.7623 0.8731
No log 23.2381 488 0.7209 0.6138 0.7209 0.8490
No log 23.3333 490 0.7463 0.5312 0.7463 0.8639
No log 23.4286 492 0.7757 0.5011 0.7757 0.8807
No log 23.5238 494 0.7540 0.5194 0.7540 0.8683
No log 23.6190 496 0.7425 0.5161 0.7425 0.8617
No log 23.7143 498 0.7400 0.5773 0.7400 0.8602
0.256 23.8095 500 0.7468 0.5363 0.7468 0.8642
0.256 23.9048 502 0.7760 0.5783 0.7760 0.8809
0.256 24.0 504 0.7826 0.5879 0.7826 0.8846
0.256 24.0952 506 0.7749 0.5732 0.7749 0.8803
0.256 24.1905 508 0.8268 0.5662 0.8268 0.9093
0.256 24.2857 510 0.8585 0.5850 0.8585 0.9266
0.256 24.3810 512 0.8338 0.5848 0.8338 0.9131
0.256 24.4762 514 0.7954 0.5405 0.7954 0.8919
0.256 24.5714 516 0.7773 0.5349 0.7773 0.8816
0.256 24.6667 518 0.7639 0.5239 0.7639 0.8740
0.256 24.7619 520 0.7797 0.5610 0.7797 0.8830
0.256 24.8571 522 0.8227 0.4948 0.8227 0.9070
0.256 24.9524 524 0.8470 0.4470 0.8470 0.9203
0.256 25.0476 526 0.8287 0.4588 0.8287 0.9104
0.256 25.1429 528 0.8103 0.4726 0.8103 0.9002
0.256 25.2381 530 0.7895 0.5236 0.7895 0.8885

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task5_organization

Finetuned
(4019)
this model