ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k5_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8447
  • Qwk: 0.6519
  • Mse: 0.8447
  • Rmse: 0.9191

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0909 2 6.9835 0.0056 6.9835 2.6426
No log 0.1818 4 4.7550 0.0 4.7550 2.1806
No log 0.2727 6 5.0364 -0.0978 5.0364 2.2442
No log 0.3636 8 3.6057 -0.0973 3.6057 1.8989
No log 0.4545 10 2.0263 0.0708 2.0263 1.4235
No log 0.5455 12 1.9269 0.1579 1.9269 1.3881
No log 0.6364 14 2.2083 0.1449 2.2083 1.4860
No log 0.7273 16 2.1982 0.1871 2.1982 1.4826
No log 0.8182 18 1.8520 0.3594 1.8520 1.3609
No log 0.9091 20 1.5491 0.3276 1.5491 1.2446
No log 1.0 22 1.5124 0.3130 1.5124 1.2298
No log 1.0909 24 1.6338 0.3130 1.6338 1.2782
No log 1.1818 26 1.4927 0.3559 1.4927 1.2218
No log 1.2727 28 1.4654 0.4252 1.4654 1.2105
No log 1.3636 30 2.1731 0.2436 2.1731 1.4742
No log 1.4545 32 2.7779 0.1446 2.7779 1.6667
No log 1.5455 34 3.0972 0.0936 3.0972 1.7599
No log 1.6364 36 2.7444 0.0848 2.7444 1.6566
No log 1.7273 38 2.1230 0.2162 2.1230 1.4571
No log 1.8182 40 1.6378 0.4593 1.6378 1.2798
No log 1.9091 42 1.2516 0.4878 1.2516 1.1188
No log 2.0 44 1.0634 0.544 1.0634 1.0312
No log 2.0909 46 1.1163 0.5970 1.1163 1.0566
No log 2.1818 48 1.5852 0.4672 1.5852 1.2590
No log 2.2727 50 2.1330 0.2516 2.1330 1.4605
No log 2.3636 52 2.0692 0.2805 2.0692 1.4385
No log 2.4545 54 1.6437 0.4384 1.6437 1.2821
No log 2.5455 56 1.1826 0.5694 1.1826 1.0875
No log 2.6364 58 1.0209 0.5821 1.0209 1.0104
No log 2.7273 60 1.2743 0.4844 1.2743 1.1289
No log 2.8182 62 1.2571 0.4806 1.2571 1.1212
No log 2.9091 64 0.9221 0.6324 0.9221 0.9603
No log 3.0 66 0.9476 0.6667 0.9476 0.9735
No log 3.0909 68 1.4582 0.5283 1.4582 1.2076
No log 3.1818 70 1.6405 0.4121 1.6405 1.2808
No log 3.2727 72 1.3786 0.5526 1.3786 1.1741
No log 3.3636 74 1.1307 0.5571 1.1307 1.0633
No log 3.4545 76 1.0490 0.5942 1.0490 1.0242
No log 3.5455 78 0.9775 0.6423 0.9775 0.9887
No log 3.6364 80 0.9697 0.5909 0.9697 0.9848
No log 3.7273 82 0.9574 0.6165 0.9574 0.9785
No log 3.8182 84 0.9349 0.6232 0.9349 0.9669
No log 3.9091 86 0.8933 0.6667 0.8933 0.9451
No log 4.0 88 1.0150 0.6707 1.0150 1.0075
No log 4.0909 90 1.2130 0.6127 1.2130 1.1014
No log 4.1818 92 1.2115 0.6228 1.2115 1.1007
No log 4.2727 94 1.1232 0.6076 1.1232 1.0598
No log 4.3636 96 1.0729 0.5972 1.0729 1.0358
No log 4.4545 98 0.9679 0.6619 0.9679 0.9838
No log 4.5455 100 0.8400 0.7273 0.8400 0.9165
No log 4.6364 102 0.7656 0.7347 0.7656 0.8750
No log 4.7273 104 0.7096 0.7568 0.7096 0.8424
No log 4.8182 106 0.7130 0.7586 0.7130 0.8444
No log 4.9091 108 0.7016 0.75 0.7016 0.8376
No log 5.0 110 0.7071 0.7273 0.7071 0.8409
No log 5.0909 112 0.7526 0.6993 0.7526 0.8675
No log 5.1818 114 0.8331 0.6761 0.8331 0.9127
No log 5.2727 116 0.8400 0.6846 0.8400 0.9165
No log 5.3636 118 0.9625 0.6667 0.9625 0.9811
No log 5.4545 120 1.0999 0.6667 1.0999 1.0488
No log 5.5455 122 1.0201 0.6364 1.0201 1.0100
No log 5.6364 124 0.9960 0.6294 0.9960 0.9980
No log 5.7273 126 1.1276 0.5303 1.1276 1.0619
No log 5.8182 128 1.1895 0.4127 1.1895 1.0906
No log 5.9091 130 0.9713 0.6377 0.9713 0.9855
No log 6.0 132 0.8628 0.7172 0.8628 0.9289
No log 6.0909 134 0.9175 0.625 0.9175 0.9579
No log 6.1818 136 0.9265 0.6081 0.9265 0.9625
No log 6.2727 138 0.8321 0.75 0.8321 0.9122
No log 6.3636 140 0.8282 0.7143 0.8282 0.9100
No log 6.4545 142 0.7235 0.76 0.7235 0.8506
No log 6.5455 144 0.7321 0.7383 0.7321 0.8556
No log 6.6364 146 0.7310 0.7619 0.7310 0.8550
No log 6.7273 148 0.9390 0.6577 0.9390 0.9690
No log 6.8182 150 1.5275 0.4969 1.5275 1.2359
No log 6.9091 152 1.5914 0.4359 1.5914 1.2615
No log 7.0 154 1.1806 0.5634 1.1806 1.0866
No log 7.0909 156 0.9173 0.6620 0.9173 0.9577
No log 7.1818 158 0.9102 0.6713 0.9102 0.9540
No log 7.2727 160 1.0485 0.6104 1.0485 1.0239
No log 7.3636 162 1.1825 0.6012 1.1825 1.0874
No log 7.4545 164 1.4073 0.4734 1.4073 1.1863
No log 7.5455 166 1.3870 0.4910 1.3870 1.1777
No log 7.6364 168 1.0688 0.6667 1.0688 1.0338
No log 7.7273 170 0.8660 0.6715 0.8660 0.9306
No log 7.8182 172 0.9740 0.6383 0.9740 0.9869
No log 7.9091 174 0.9817 0.6479 0.9817 0.9908
No log 8.0 176 0.9095 0.6857 0.9095 0.9537
No log 8.0909 178 0.8751 0.6993 0.8751 0.9354
No log 8.1818 180 0.8358 0.7133 0.8358 0.9142
No log 8.2727 182 0.8126 0.6765 0.8126 0.9014
No log 8.3636 184 0.8151 0.6471 0.8151 0.9028
No log 8.4545 186 0.8496 0.6061 0.8496 0.9218
No log 8.5455 188 0.8897 0.5920 0.8897 0.9432
No log 8.6364 190 0.9017 0.5891 0.9017 0.9496
No log 8.7273 192 0.8653 0.6074 0.8653 0.9302
No log 8.8182 194 0.8090 0.7092 0.8090 0.8994
No log 8.9091 196 0.7783 0.7042 0.7783 0.8822
No log 9.0 198 0.7822 0.7297 0.7822 0.8844
No log 9.0909 200 0.7478 0.7368 0.7478 0.8648
No log 9.1818 202 0.6939 0.7383 0.6939 0.8330
No log 9.2727 204 0.7293 0.6667 0.7293 0.8540
No log 9.3636 206 0.8160 0.6377 0.8160 0.9033
No log 9.4545 208 0.8733 0.6471 0.8733 0.9345
No log 9.5455 210 0.9505 0.5957 0.9505 0.9749
No log 9.6364 212 1.0093 0.6056 1.0093 1.0046
No log 9.7273 214 1.0039 0.5441 1.0039 1.0019
No log 9.8182 216 0.9703 0.5538 0.9703 0.9850
No log 9.9091 218 0.9739 0.5397 0.9739 0.9869
No log 10.0 220 1.0077 0.5469 1.0077 1.0039
No log 10.0909 222 0.8938 0.5649 0.8938 0.9454
No log 10.1818 224 0.7583 0.7286 0.7583 0.8708
No log 10.2727 226 0.7102 0.7 0.7102 0.8427
No log 10.3636 228 0.6614 0.7310 0.6614 0.8133
No log 10.4545 230 0.6770 0.7133 0.6770 0.8228
No log 10.5455 232 0.7580 0.7172 0.7580 0.8706
No log 10.6364 234 0.7564 0.6853 0.7564 0.8697
No log 10.7273 236 0.7783 0.7222 0.7783 0.8822
No log 10.8182 238 0.9561 0.5891 0.9561 0.9778
No log 10.9091 240 0.8893 0.5692 0.8893 0.9430
No log 11.0 242 0.6791 0.7383 0.6791 0.8241
No log 11.0909 244 0.7216 0.72 0.7216 0.8495
No log 11.1818 246 0.8570 0.6974 0.8570 0.9258
No log 11.2727 248 0.7892 0.7162 0.7892 0.8884
No log 11.3636 250 0.7404 0.6950 0.7404 0.8605
No log 11.4545 252 0.7557 0.6803 0.7557 0.8693
No log 11.5455 254 0.7447 0.6986 0.7447 0.8630
No log 11.6364 256 0.7536 0.7260 0.7536 0.8681
No log 11.7273 258 0.7626 0.7222 0.7626 0.8733
No log 11.8182 260 0.7918 0.7042 0.7918 0.8899
No log 11.9091 262 0.8005 0.6857 0.8005 0.8947
No log 12.0 264 0.8234 0.6522 0.8234 0.9074
No log 12.0909 266 0.8500 0.6714 0.8500 0.9219
No log 12.1818 268 0.8646 0.7042 0.8646 0.9299
No log 12.2727 270 0.7733 0.6901 0.7733 0.8794
No log 12.3636 272 0.7110 0.6901 0.7110 0.8432
No log 12.4545 274 0.7223 0.7183 0.7223 0.8499
No log 12.5455 276 0.6917 0.7133 0.6917 0.8317
No log 12.6364 278 0.7029 0.7183 0.7029 0.8384
No log 12.7273 280 0.7836 0.6986 0.7836 0.8852
No log 12.8182 282 0.8623 0.6939 0.8623 0.9286
No log 12.9091 284 0.8333 0.6761 0.8333 0.9128
No log 13.0 286 0.7834 0.6667 0.7834 0.8851
No log 13.0909 288 0.7381 0.6906 0.7381 0.8591
No log 13.1818 290 0.7322 0.7376 0.7322 0.8557
No log 13.2727 292 0.7168 0.7092 0.7168 0.8467
No log 13.3636 294 0.7431 0.7083 0.7431 0.8621
No log 13.4545 296 0.8190 0.7027 0.8190 0.9050
No log 13.5455 298 0.8424 0.6806 0.8424 0.9178
No log 13.6364 300 0.7868 0.7123 0.7868 0.8870
No log 13.7273 302 0.7634 0.7260 0.7634 0.8738
No log 13.8182 304 0.7724 0.7260 0.7724 0.8789
No log 13.9091 306 0.7703 0.7273 0.7703 0.8777
No log 14.0 308 0.7734 0.6950 0.7734 0.8794
No log 14.0909 310 0.7654 0.6571 0.7654 0.8749
No log 14.1818 312 0.7676 0.7092 0.7676 0.8761
No log 14.2727 314 0.8231 0.6950 0.8231 0.9072
No log 14.3636 316 0.8691 0.6714 0.8691 0.9322
No log 14.4545 318 0.8983 0.6714 0.8983 0.9478
No log 14.5455 320 0.8615 0.7 0.8615 0.9282
No log 14.6364 322 0.8347 0.7183 0.8347 0.9136
No log 14.7273 324 0.8119 0.6475 0.8119 0.9010
No log 14.8182 326 0.7959 0.6569 0.7959 0.8922
No log 14.9091 328 0.7901 0.6619 0.7901 0.8889
No log 15.0 330 0.7542 0.6620 0.7542 0.8685
No log 15.0909 332 0.7736 0.6761 0.7736 0.8796
No log 15.1818 334 0.7924 0.6761 0.7924 0.8902
No log 15.2727 336 0.8036 0.6761 0.8036 0.8965
No log 15.3636 338 0.7768 0.6714 0.7768 0.8813
No log 15.4545 340 0.7791 0.6619 0.7791 0.8826
No log 15.5455 342 0.7772 0.6667 0.7772 0.8816
No log 15.6364 344 0.7928 0.6667 0.7928 0.8904
No log 15.7273 346 0.8067 0.6763 0.8067 0.8982
No log 15.8182 348 0.8227 0.6857 0.8227 0.9070
No log 15.9091 350 0.8209 0.6950 0.8209 0.9060
No log 16.0 352 0.8146 0.6763 0.8146 0.9026
No log 16.0909 354 0.7965 0.6569 0.7965 0.8925
No log 16.1818 356 0.8184 0.7194 0.8184 0.9047
No log 16.2727 358 0.8188 0.7194 0.8188 0.9049
No log 16.3636 360 0.7802 0.7007 0.7802 0.8833
No log 16.4545 362 0.7696 0.6815 0.7696 0.8773
No log 16.5455 364 0.7831 0.6763 0.7831 0.8849
No log 16.6364 366 0.7510 0.6763 0.7510 0.8666
No log 16.7273 368 0.7036 0.7234 0.7036 0.8388
No log 16.8182 370 0.7099 0.6950 0.7099 0.8425
No log 16.9091 372 0.7149 0.6901 0.7149 0.8455
No log 17.0 374 0.7074 0.6906 0.7074 0.8411
No log 17.0909 376 0.7577 0.6763 0.7577 0.8704
No log 17.1818 378 0.8195 0.6667 0.8195 0.9053
No log 17.2727 380 0.8345 0.6716 0.8345 0.9135
No log 17.3636 382 0.8223 0.6716 0.8223 0.9068
No log 17.4545 384 0.7715 0.6812 0.7715 0.8784
No log 17.5455 386 0.7488 0.7143 0.7488 0.8653
No log 17.6364 388 0.7371 0.6857 0.7371 0.8586
No log 17.7273 390 0.7135 0.6809 0.7135 0.8447
No log 17.8182 392 0.6906 0.7083 0.6906 0.8310
No log 17.9091 394 0.6807 0.7083 0.6807 0.8250
No log 18.0 396 0.6928 0.6901 0.6928 0.8323
No log 18.0909 398 0.7063 0.7286 0.7063 0.8404
No log 18.1818 400 0.7307 0.7246 0.7307 0.8548
No log 18.2727 402 0.7721 0.7059 0.7721 0.8787
No log 18.3636 404 0.7922 0.6716 0.7922 0.8900
No log 18.4545 406 0.7884 0.6815 0.7884 0.8879
No log 18.5455 408 0.7936 0.6716 0.7936 0.8908
No log 18.6364 410 0.7723 0.7059 0.7723 0.8788
No log 18.7273 412 0.7492 0.6812 0.7492 0.8655
No log 18.8182 414 0.7496 0.7 0.7496 0.8658
No log 18.9091 416 0.7683 0.6950 0.7683 0.8765
No log 19.0 418 0.8201 0.6857 0.8201 0.9056
No log 19.0909 420 0.8191 0.6857 0.8191 0.9050
No log 19.1818 422 0.7955 0.6812 0.7955 0.8919
No log 19.2727 424 0.8045 0.7007 0.8045 0.8969
No log 19.3636 426 0.8160 0.7007 0.8160 0.9033
No log 19.4545 428 0.8283 0.7007 0.8283 0.9101
No log 19.5455 430 0.8378 0.7007 0.8378 0.9153
No log 19.6364 432 0.8351 0.7007 0.8351 0.9138
No log 19.7273 434 0.8039 0.6906 0.8039 0.8966
No log 19.8182 436 0.7788 0.7 0.7788 0.8825
No log 19.9091 438 0.7607 0.6957 0.7607 0.8722
No log 20.0 440 0.7796 0.7234 0.7796 0.8829
No log 20.0909 442 0.7844 0.7234 0.7844 0.8857
No log 20.1818 444 0.7735 0.7050 0.7735 0.8795
No log 20.2727 446 0.7800 0.6667 0.7800 0.8832
No log 20.3636 448 0.8326 0.6522 0.8326 0.9125
No log 20.4545 450 0.8289 0.6619 0.8289 0.9104
No log 20.5455 452 0.7690 0.6809 0.7690 0.8770
No log 20.6364 454 0.7691 0.6993 0.7691 0.8770
No log 20.7273 456 0.8953 0.6522 0.8953 0.9462
No log 20.8182 458 1.0292 0.6 1.0292 1.0145
No log 20.9091 460 1.0586 0.5899 1.0586 1.0289
No log 21.0 462 0.9657 0.6176 0.9657 0.9827
No log 21.0909 464 0.8692 0.6423 0.8692 0.9323
No log 21.1818 466 0.8421 0.6716 0.8421 0.9177
No log 21.2727 468 0.8754 0.6667 0.8754 0.9356
No log 21.3636 470 0.8849 0.6667 0.8849 0.9407
No log 21.4545 472 0.9033 0.6364 0.9033 0.9504
No log 21.5455 474 0.9287 0.6260 0.9287 0.9637
No log 21.6364 476 0.9505 0.6154 0.9505 0.9750
No log 21.7273 478 0.8992 0.6316 0.8992 0.9483
No log 21.8182 480 0.8072 0.7007 0.8072 0.8985
No log 21.9091 482 0.7318 0.7050 0.7318 0.8555
No log 22.0 484 0.6989 0.7361 0.6989 0.8360
No log 22.0909 486 0.6833 0.7361 0.6833 0.8266
No log 22.1818 488 0.6906 0.7361 0.6906 0.8310
No log 22.2727 490 0.6972 0.7310 0.6972 0.8350
No log 22.3636 492 0.6996 0.7534 0.6996 0.8364
No log 22.4545 494 0.7268 0.7483 0.7268 0.8525
No log 22.5455 496 0.7685 0.7133 0.7685 0.8766
No log 22.6364 498 0.8125 0.6809 0.8125 0.9014
0.3761 22.7273 500 0.8225 0.6809 0.8225 0.9069
0.3761 22.8182 502 0.7949 0.7183 0.7949 0.8916
0.3761 22.9091 504 0.7788 0.6950 0.7788 0.8825
0.3761 23.0 506 0.7730 0.6857 0.7730 0.8792
0.3761 23.0909 508 0.7581 0.7 0.7581 0.8707
0.3761 23.1818 510 0.7432 0.7092 0.7432 0.8621
0.3761 23.2727 512 0.7090 0.7092 0.7090 0.8420
0.3761 23.3636 514 0.6861 0.7361 0.6861 0.8283
0.3761 23.4545 516 0.6932 0.7361 0.6932 0.8326
0.3761 23.5455 518 0.6995 0.7361 0.6995 0.8363
0.3761 23.6364 520 0.7301 0.7042 0.7301 0.8544
0.3761 23.7273 522 0.7653 0.7183 0.7653 0.8748
0.3761 23.8182 524 0.7722 0.7183 0.7722 0.8788
0.3761 23.9091 526 0.7928 0.6950 0.7928 0.8904
0.3761 24.0 528 0.8384 0.7 0.8384 0.9156
0.3761 24.0909 530 0.9017 0.6475 0.9017 0.9496
0.3761 24.1818 532 0.9052 0.6475 0.9052 0.9514
0.3761 24.2727 534 0.8898 0.6522 0.8898 0.9433
0.3761 24.3636 536 0.8477 0.6715 0.8477 0.9207
0.3761 24.4545 538 0.8138 0.6906 0.8138 0.9021
0.3761 24.5455 540 0.8080 0.6906 0.8080 0.8989
0.3761 24.6364 542 0.8072 0.6906 0.8072 0.8984
0.3761 24.7273 544 0.8105 0.6812 0.8105 0.9003
0.3761 24.8182 546 0.8272 0.6861 0.8272 0.9095
0.3761 24.9091 548 0.8505 0.7042 0.8505 0.9222
0.3761 25.0 550 0.8664 0.6906 0.8664 0.9308
0.3761 25.0909 552 0.8594 0.6906 0.8594 0.9270
0.3761 25.1818 554 0.8309 0.6861 0.8309 0.9115
0.3761 25.2727 556 0.8308 0.6569 0.8308 0.9115
0.3761 25.3636 558 0.8580 0.6423 0.8580 0.9263
0.3761 25.4545 560 0.8769 0.6418 0.8769 0.9364
0.3761 25.5455 562 0.8733 0.6519 0.8733 0.9345
0.3761 25.6364 564 0.8447 0.6519 0.8447 0.9191

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k5_task1_organization

Finetuned
(4019)
this model