ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k20_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8250
  • Qwk: 0.5839
  • Mse: 0.8250
  • Rmse: 0.9083

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0211 2 6.7085 -0.0058 6.7085 2.5901
No log 0.0421 4 4.3923 0.0 4.3923 2.0958
No log 0.0632 6 3.8760 -0.0597 3.8760 1.9688
No log 0.0842 8 3.7710 -0.0808 3.7710 1.9419
No log 0.1053 10 2.3712 0.0576 2.3712 1.5399
No log 0.1263 12 2.2859 0.0567 2.2859 1.5119
No log 0.1474 14 2.6803 0.0649 2.6803 1.6372
No log 0.1684 16 2.6750 0.0764 2.6750 1.6355
No log 0.1895 18 2.3076 0.125 2.3076 1.5191
No log 0.2105 20 2.0487 0.2137 2.0487 1.4313
No log 0.2316 22 2.0204 0.3308 2.0204 1.4214
No log 0.2526 24 2.0691 0.3014 2.0691 1.4384
No log 0.2737 26 2.2325 0.1899 2.2325 1.4942
No log 0.2947 28 2.1731 0.2452 2.1731 1.4742
No log 0.3158 30 1.7083 0.4503 1.7083 1.3070
No log 0.3368 32 1.1956 0.5942 1.1956 1.0934
No log 0.3579 34 1.2564 0.5298 1.2564 1.1209
No log 0.3789 36 1.8935 0.4483 1.8935 1.3760
No log 0.4 38 2.6962 0.2843 2.6962 1.6420
No log 0.4211 40 2.9775 0.2449 2.9775 1.7256
No log 0.4421 42 2.3310 0.2125 2.3310 1.5268
No log 0.4632 44 1.4082 0.4853 1.4082 1.1867
No log 0.4842 46 1.1630 0.5455 1.1630 1.0784
No log 0.5053 48 1.0909 0.6074 1.0909 1.0445
No log 0.5263 50 1.0658 0.6241 1.0658 1.0324
No log 0.5474 52 1.1206 0.6111 1.1206 1.0586
No log 0.5684 54 0.9190 0.6950 0.9190 0.9587
No log 0.5895 56 0.8780 0.6809 0.8780 0.9370
No log 0.6105 58 0.9478 0.6761 0.9478 0.9736
No log 0.6316 60 1.3475 0.5419 1.3475 1.1608
No log 0.6526 62 1.3406 0.5316 1.3406 1.1578
No log 0.6737 64 1.1821 0.5802 1.1821 1.0872
No log 0.6947 66 0.8267 0.7273 0.8267 0.9092
No log 0.7158 68 0.8813 0.7260 0.8813 0.9388
No log 0.7368 70 0.9476 0.6879 0.9476 0.9735
No log 0.7579 72 1.4897 0.5455 1.4897 1.2206
No log 0.7789 74 1.7566 0.4309 1.7566 1.3254
No log 0.8 76 1.1753 0.6289 1.1753 1.0841
No log 0.8211 78 0.6857 0.7815 0.6857 0.8281
No log 0.8421 80 0.6575 0.7895 0.6575 0.8109
No log 0.8632 82 0.7298 0.75 0.7298 0.8543
No log 0.8842 84 0.9554 0.6905 0.9554 0.9774
No log 0.9053 86 1.1751 0.6703 1.1751 1.0840
No log 0.9263 88 1.0072 0.7018 1.0072 1.0036
No log 0.9474 90 0.7698 0.7342 0.7698 0.8774
No log 0.9684 92 0.7519 0.6806 0.7519 0.8671
No log 0.9895 94 0.9463 0.6569 0.9463 0.9728
No log 1.0105 96 1.0611 0.5538 1.0611 1.0301
No log 1.0316 98 0.9447 0.5564 0.9447 0.9719
No log 1.0526 100 0.9368 0.6286 0.9368 0.9679
No log 1.0737 102 1.2629 0.6154 1.2629 1.1238
No log 1.0947 104 1.4679 0.5568 1.4679 1.2115
No log 1.1158 106 1.2806 0.6303 1.2806 1.1317
No log 1.1368 108 1.1117 0.6144 1.1117 1.0544
No log 1.1579 110 1.1558 0.6076 1.1558 1.0751
No log 1.1789 112 1.4003 0.5795 1.4003 1.1833
No log 1.2 114 1.2457 0.6136 1.2457 1.1161
No log 1.2211 116 1.1441 0.6592 1.1441 1.0696
No log 1.2421 118 0.8906 0.6792 0.8906 0.9437
No log 1.2632 120 0.8526 0.6980 0.8526 0.9234
No log 1.2842 122 0.9436 0.6621 0.9436 0.9714
No log 1.3053 124 0.8503 0.7075 0.8503 0.9221
No log 1.3263 126 0.7702 0.7027 0.7702 0.8776
No log 1.3474 128 0.8195 0.6986 0.8195 0.9053
No log 1.3684 130 0.7095 0.7550 0.7095 0.8423
No log 1.3895 132 0.6144 0.7922 0.6144 0.7838
No log 1.4105 134 0.6062 0.7333 0.6062 0.7786
No log 1.4316 136 0.6068 0.76 0.6068 0.7790
No log 1.4526 138 0.6239 0.7922 0.6239 0.7899
No log 1.4737 140 0.6960 0.7950 0.6960 0.8343
No log 1.4947 142 0.6255 0.7871 0.6255 0.7909
No log 1.5158 144 0.6340 0.7448 0.6340 0.7963
No log 1.5368 146 0.6482 0.7792 0.6482 0.8051
No log 1.5579 148 0.7805 0.7436 0.7805 0.8835
No log 1.5789 150 0.7494 0.7826 0.7494 0.8657
No log 1.6 152 0.8466 0.7215 0.8466 0.9201
No log 1.6211 154 0.9351 0.6744 0.9351 0.9670
No log 1.6421 156 0.8809 0.7135 0.8809 0.9386
No log 1.6632 158 0.7055 0.8263 0.7055 0.8399
No log 1.6842 160 0.6651 0.8221 0.6651 0.8155
No log 1.7053 162 0.7333 0.8098 0.7333 0.8563
No log 1.7263 164 1.0160 0.6590 1.0160 1.0079
No log 1.7474 166 1.1017 0.6333 1.1017 1.0496
No log 1.7684 168 0.7692 0.7692 0.7692 0.8770
No log 1.7895 170 0.6246 0.7871 0.6246 0.7903
No log 1.8105 172 0.7037 0.7595 0.7037 0.8389
No log 1.8316 174 0.6289 0.7643 0.6289 0.7930
No log 1.8526 176 0.6494 0.8090 0.6494 0.8059
No log 1.8737 178 0.8667 0.7640 0.8667 0.9310
No log 1.8947 180 0.7988 0.7571 0.7988 0.8938
No log 1.9158 182 0.6130 0.8228 0.6130 0.7829
No log 1.9368 184 0.6864 0.7586 0.6864 0.8285
No log 1.9579 186 0.7612 0.7299 0.7612 0.8725
No log 1.9789 188 0.8460 0.5846 0.8460 0.9198
No log 2.0 190 0.9706 0.5942 0.9706 0.9852
No log 2.0211 192 0.9400 0.6099 0.9400 0.9695
No log 2.0421 194 0.6958 0.7517 0.6958 0.8341
No log 2.0632 196 0.6330 0.7451 0.6330 0.7956
No log 2.0842 198 0.5906 0.7702 0.5906 0.7685
No log 2.1053 200 0.5698 0.7875 0.5698 0.7548
No log 2.1263 202 0.5781 0.8176 0.5781 0.7603
No log 2.1474 204 0.6626 0.8 0.6626 0.8140
No log 2.1684 206 1.0182 0.6489 1.0182 1.0091
No log 2.1895 208 1.7271 0.6019 1.7271 1.3142
No log 2.2105 210 2.0776 0.5023 2.0776 1.4414
No log 2.2316 212 1.5579 0.6082 1.5579 1.2482
No log 2.2526 214 0.8891 0.6824 0.8891 0.9429
No log 2.2737 216 0.7194 0.7517 0.7194 0.8482
No log 2.2947 218 0.7010 0.7703 0.7010 0.8373
No log 2.3158 220 0.7263 0.7368 0.7263 0.8522
No log 2.3368 222 0.7488 0.6846 0.7488 0.8654
No log 2.3579 224 0.7797 0.6712 0.7797 0.8830
No log 2.3789 226 0.7572 0.6857 0.7572 0.8702
No log 2.4 228 0.7638 0.6857 0.7638 0.8740
No log 2.4211 230 0.7376 0.6993 0.7376 0.8588
No log 2.4421 232 0.7089 0.7417 0.7089 0.8420
No log 2.4632 234 0.7541 0.7425 0.7541 0.8684
No log 2.4842 236 0.7309 0.7910 0.7309 0.8549
No log 2.5053 238 0.6367 0.7514 0.6367 0.7979
No log 2.5263 240 0.6420 0.7453 0.6420 0.8012
No log 2.5474 242 0.6139 0.7758 0.6139 0.7835
No log 2.5684 244 0.6295 0.8221 0.6295 0.7934
No log 2.5895 246 0.7371 0.7636 0.7371 0.8585
No log 2.6105 248 0.7446 0.7547 0.7446 0.8629
No log 2.6316 250 0.7101 0.7651 0.7101 0.8427
No log 2.6526 252 0.6888 0.7651 0.6888 0.8299
No log 2.6737 254 0.6678 0.7651 0.6678 0.8172
No log 2.6947 256 0.7226 0.7778 0.7226 0.8501
No log 2.7158 258 0.8305 0.7349 0.8305 0.9113
No log 2.7368 260 0.7952 0.7602 0.7952 0.8917
No log 2.7579 262 0.8939 0.7151 0.8939 0.9455
No log 2.7789 264 0.9057 0.7151 0.9057 0.9517
No log 2.8 266 0.8374 0.7345 0.8374 0.9151
No log 2.8211 268 0.8090 0.7514 0.8090 0.8994
No log 2.8421 270 0.7275 0.7953 0.7275 0.8530
No log 2.8632 272 0.6204 0.8313 0.6204 0.7877
No log 2.8842 274 0.6262 0.8125 0.6262 0.7913
No log 2.9053 276 0.7829 0.7089 0.7829 0.8848
No log 2.9263 278 1.0735 0.6918 1.0735 1.0361
No log 2.9474 280 1.0504 0.6795 1.0504 1.0249
No log 2.9684 282 0.8457 0.625 0.8457 0.9196
No log 2.9895 284 0.7131 0.7483 0.7131 0.8445
No log 3.0105 286 0.6662 0.7815 0.6662 0.8162
No log 3.0316 288 0.7009 0.7730 0.7009 0.8372
No log 3.0526 290 0.7915 0.7647 0.7915 0.8897
No log 3.0737 292 0.7443 0.7574 0.7443 0.8628
No log 3.0947 294 0.6511 0.7950 0.6511 0.8069
No log 3.1158 296 0.6714 0.7582 0.6714 0.8194
No log 3.1368 298 0.6764 0.7815 0.6764 0.8225
No log 3.1579 300 0.7076 0.7799 0.7076 0.8412
No log 3.1789 302 0.8647 0.6829 0.8647 0.9299
No log 3.2 304 0.8696 0.6905 0.8696 0.9325
No log 3.2211 306 0.7225 0.7375 0.7225 0.8500
No log 3.2421 308 0.6378 0.7950 0.6378 0.7986
No log 3.2632 310 0.6722 0.7853 0.6722 0.8199
No log 3.2842 312 0.8389 0.6753 0.8389 0.9159
No log 3.3053 314 0.9165 0.6835 0.9165 0.9574
No log 3.3263 316 0.8551 0.6623 0.8551 0.9247
No log 3.3474 318 0.6640 0.7925 0.6640 0.8149
No log 3.3684 320 0.6572 0.7799 0.6572 0.8107
No log 3.3895 322 0.7068 0.7811 0.7068 0.8407
No log 3.4105 324 0.7102 0.7955 0.7102 0.8428
No log 3.4316 326 0.6716 0.8023 0.6716 0.8195
No log 3.4526 328 0.6099 0.7977 0.6099 0.7810
No log 3.4737 330 0.5753 0.8024 0.5753 0.7585
No log 3.4947 332 0.6013 0.7925 0.6013 0.7755
No log 3.5158 334 0.6690 0.7417 0.6690 0.8179
No log 3.5368 336 0.7516 0.7114 0.7516 0.8669
No log 3.5579 338 0.7786 0.7170 0.7786 0.8824
No log 3.5789 340 0.8434 0.7219 0.8434 0.9184
No log 3.6 342 0.9837 0.6893 0.9837 0.9918
No log 3.6211 344 0.8317 0.7514 0.8317 0.9120
No log 3.6421 346 0.6343 0.8208 0.6343 0.7964
No log 3.6632 348 0.5839 0.8049 0.5839 0.7641
No log 3.6842 350 0.6044 0.7975 0.6044 0.7774
No log 3.7053 352 0.6258 0.8313 0.6258 0.7911
No log 3.7263 354 0.6722 0.7831 0.6722 0.8198
No log 3.7474 356 0.7034 0.7619 0.7034 0.8387
No log 3.7684 358 0.6452 0.7826 0.6452 0.8033
No log 3.7895 360 0.6764 0.775 0.6764 0.8225
No log 3.8105 362 0.6564 0.7771 0.6564 0.8102
No log 3.8316 364 0.6611 0.7826 0.6611 0.8131
No log 3.8526 366 0.6175 0.8242 0.6175 0.7858
No log 3.8737 368 0.5962 0.8402 0.5962 0.7722
No log 3.8947 370 0.5916 0.8214 0.5916 0.7692
No log 3.9158 372 0.5527 0.8171 0.5527 0.7435
No log 3.9368 374 0.5566 0.8075 0.5566 0.7461
No log 3.9579 376 0.5644 0.8125 0.5644 0.7513
No log 3.9789 378 0.6727 0.7904 0.6727 0.8202
No log 4.0 380 0.7162 0.7439 0.7162 0.8463
No log 4.0211 382 0.6747 0.7692 0.6747 0.8214
No log 4.0421 384 0.7270 0.7439 0.7270 0.8526
No log 4.0632 386 0.6756 0.7673 0.6756 0.8219
No log 4.0842 388 0.5905 0.7975 0.5905 0.7684
No log 4.1053 390 0.5555 0.8025 0.5555 0.7453
No log 4.1263 392 0.5473 0.8075 0.5473 0.7398
No log 4.1474 394 0.5564 0.8402 0.5564 0.7459
No log 4.1684 396 0.5962 0.8229 0.5962 0.7722
No log 4.1895 398 0.5808 0.8409 0.5808 0.7621
No log 4.2105 400 0.6094 0.8333 0.6094 0.7807
No log 4.2316 402 0.7874 0.7444 0.7874 0.8874
No log 4.2526 404 0.8364 0.7444 0.8364 0.9145
No log 4.2737 406 0.7362 0.7910 0.7362 0.8580
No log 4.2947 408 0.6395 0.825 0.6395 0.7997
No log 4.3158 410 0.6648 0.7871 0.6648 0.8153
No log 4.3368 412 0.6718 0.8176 0.6718 0.8196
No log 4.3579 414 0.6892 0.7702 0.6892 0.8302
No log 4.3789 416 0.6603 0.7879 0.6603 0.8126
No log 4.4 418 0.6078 0.8121 0.6078 0.7796
No log 4.4211 420 0.5371 0.8101 0.5371 0.7329
No log 4.4421 422 0.5346 0.8 0.5346 0.7312
No log 4.4632 424 0.5312 0.7821 0.5312 0.7288
No log 4.4842 426 0.5273 0.7975 0.5273 0.7261
No log 4.5053 428 0.5778 0.7975 0.5778 0.7602
No log 4.5263 430 0.6971 0.7701 0.6971 0.8349
No log 4.5474 432 0.7795 0.7571 0.7795 0.8829
No log 4.5684 434 0.7325 0.7753 0.7325 0.8559
No log 4.5895 436 0.6473 0.8140 0.6473 0.8046
No log 4.6105 438 0.6318 0.8121 0.6318 0.7949
No log 4.6316 440 0.6295 0.7742 0.6295 0.7934
No log 4.6526 442 0.6082 0.7821 0.6082 0.7799
No log 4.6737 444 0.5688 0.8153 0.5688 0.7542
No log 4.6947 446 0.5798 0.7975 0.5798 0.7614
No log 4.7158 448 0.6148 0.8176 0.6148 0.7841
No log 4.7368 450 0.7174 0.7329 0.7174 0.8470
No log 4.7579 452 0.7640 0.6918 0.7640 0.8741
No log 4.7789 454 0.6787 0.7564 0.6787 0.8238
No log 4.8 456 0.6246 0.7925 0.6246 0.7903
No log 4.8211 458 0.6401 0.7758 0.6401 0.8001
No log 4.8421 460 0.7341 0.7381 0.7341 0.8568
No log 4.8632 462 0.7749 0.7399 0.7749 0.8803
No log 4.8842 464 0.7599 0.7412 0.7599 0.8717
No log 4.9053 466 0.8404 0.6988 0.8404 0.9167
No log 4.9263 468 0.8340 0.6710 0.8340 0.9132
No log 4.9474 470 0.7026 0.7114 0.7026 0.8382
No log 4.9684 472 0.5727 0.7755 0.5727 0.7568
No log 4.9895 474 0.5742 0.7785 0.5742 0.7578
No log 5.0105 476 0.7232 0.7811 0.7232 0.8504
No log 5.0316 478 0.9878 0.6782 0.9878 0.9939
No log 5.0526 480 0.9738 0.6821 0.9738 0.9868
No log 5.0737 482 0.7686 0.7586 0.7686 0.8767
No log 5.0947 484 0.6150 0.7843 0.6150 0.7842
No log 5.1158 486 0.6046 0.7974 0.6046 0.7776
No log 5.1368 488 0.6274 0.7898 0.6274 0.7921
No log 5.1579 490 0.6807 0.7722 0.6807 0.8250
No log 5.1789 492 0.7421 0.7089 0.7421 0.8614
No log 5.2 494 0.6989 0.7516 0.6989 0.8360
No log 5.2211 496 0.7417 0.7013 0.7417 0.8612
No log 5.2421 498 0.7685 0.6846 0.7685 0.8766
0.3859 5.2632 500 0.7965 0.6846 0.7965 0.8925
0.3859 5.2842 502 0.7609 0.6759 0.7609 0.8723
0.3859 5.3053 504 0.8143 0.6486 0.8143 0.9024
0.3859 5.3263 506 0.7924 0.6883 0.7924 0.8902
0.3859 5.3474 508 0.7242 0.7578 0.7242 0.8510
0.3859 5.3684 510 0.7197 0.7831 0.7197 0.8484
0.3859 5.3895 512 0.6957 0.7730 0.6957 0.8341
0.3859 5.4105 514 0.7138 0.75 0.7138 0.8449
0.3859 5.4316 516 0.8489 0.7 0.8489 0.9214
0.3859 5.4526 518 0.8271 0.6923 0.8271 0.9095
0.3859 5.4737 520 0.7297 0.6944 0.7297 0.8542
0.3859 5.4947 522 0.6734 0.7448 0.6734 0.8206
0.3859 5.5158 524 0.6560 0.7568 0.6560 0.8099
0.3859 5.5368 526 0.6789 0.7673 0.6789 0.8239
0.3859 5.5579 528 0.7456 0.7152 0.7456 0.8635
0.3859 5.5789 530 0.6817 0.7470 0.6817 0.8256
0.3859 5.6 532 0.5739 0.7712 0.5739 0.7576
0.3859 5.6211 534 0.5685 0.7733 0.5685 0.7540
0.3859 5.6421 536 0.6449 0.7785 0.6449 0.8030
0.3859 5.6632 538 0.6158 0.7568 0.6158 0.7847
0.3859 5.6842 540 0.6755 0.7211 0.6755 0.8219
0.3859 5.7053 542 0.8784 0.675 0.8784 0.9372
0.3859 5.7263 544 1.0076 0.6707 1.0076 1.0038
0.3859 5.7474 546 0.8951 0.6667 0.8951 0.9461
0.3859 5.7684 548 0.8031 0.6443 0.8031 0.8961
0.3859 5.7895 550 0.7331 0.7133 0.7331 0.8562
0.3859 5.8105 552 0.7057 0.7534 0.7057 0.8400
0.3859 5.8316 554 0.6800 0.7534 0.6800 0.8246
0.3859 5.8526 556 0.6683 0.7534 0.6683 0.8175
0.3859 5.8737 558 0.6702 0.7310 0.6702 0.8186
0.3859 5.8947 560 0.6765 0.7383 0.6765 0.8225
0.3859 5.9158 562 0.6703 0.7467 0.6703 0.8187
0.3859 5.9368 564 0.6915 0.7702 0.6915 0.8315
0.3859 5.9579 566 0.7388 0.7229 0.7388 0.8595
0.3859 5.9789 568 0.8549 0.6747 0.8549 0.9246
0.3859 6.0 570 1.0284 0.6626 1.0284 1.0141
0.3859 6.0211 572 1.0392 0.65 1.0392 1.0194
0.3859 6.0421 574 0.9293 0.5899 0.9293 0.9640
0.3859 6.0632 576 0.8250 0.5839 0.8250 0.9083

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k20_task1_organization

Finetuned
(4023)
this model