ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k15_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6748
  • Qwk: 0.7015
  • Mse: 0.6748
  • Rmse: 0.8214

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0282 2 6.8284 0.0239 6.8284 2.6131
No log 0.0563 4 4.5847 0.0772 4.5847 2.1412
No log 0.0845 6 3.2874 -0.0117 3.2874 1.8131
No log 0.1127 8 2.5326 0.0580 2.5326 1.5914
No log 0.1408 10 2.1036 0.0847 2.1036 1.4504
No log 0.1690 12 1.7942 0.1273 1.7942 1.3395
No log 0.1972 14 2.2051 0.1471 2.2051 1.4850
No log 0.2254 16 1.8886 0.2564 1.8886 1.3743
No log 0.2535 18 1.9081 0.3279 1.9081 1.3814
No log 0.2817 20 1.6213 0.2281 1.6213 1.2733
No log 0.3099 22 1.6032 0.2881 1.6032 1.2662
No log 0.3380 24 1.4841 0.3590 1.4841 1.2182
No log 0.3662 26 1.3783 0.3009 1.3783 1.1740
No log 0.3944 28 1.3763 0.3833 1.3763 1.1732
No log 0.4225 30 1.3182 0.4733 1.3182 1.1481
No log 0.4507 32 1.2517 0.5 1.2517 1.1188
No log 0.4789 34 1.3285 0.4733 1.3285 1.1526
No log 0.5070 36 1.4489 0.4898 1.4489 1.2037
No log 0.5352 38 1.4877 0.4898 1.4877 1.2197
No log 0.5634 40 1.4065 0.5103 1.4065 1.1860
No log 0.5915 42 1.5065 0.4762 1.5065 1.2274
No log 0.6197 44 1.9121 0.4444 1.9121 1.3828
No log 0.6479 46 2.1736 0.3017 2.1736 1.4743
No log 0.6761 48 2.0598 0.3468 2.0598 1.4352
No log 0.7042 50 1.6580 0.4937 1.6580 1.2876
No log 0.7324 52 1.3973 0.5067 1.3973 1.1821
No log 0.7606 54 1.3256 0.5147 1.3256 1.1513
No log 0.7887 56 1.0689 0.6014 1.0689 1.0339
No log 0.8169 58 1.0468 0.6323 1.0468 1.0231
No log 0.8451 60 0.9586 0.6531 0.9586 0.9791
No log 0.8732 62 0.9158 0.6763 0.9158 0.9570
No log 0.9014 64 1.2193 0.4677 1.2193 1.1042
No log 0.9296 66 1.2355 0.5079 1.2355 1.1115
No log 0.9577 68 0.7880 0.7123 0.7880 0.8877
No log 0.9859 70 1.4157 0.5780 1.4157 1.1898
No log 1.0141 72 2.2375 0.4476 2.2375 1.4958
No log 1.0423 74 1.7493 0.5 1.7493 1.3226
No log 1.0704 76 0.8839 0.6538 0.8839 0.9401
No log 1.0986 78 0.7321 0.7483 0.7321 0.8556
No log 1.1268 80 0.7393 0.7234 0.7393 0.8598
No log 1.1549 82 0.7579 0.7092 0.7579 0.8706
No log 1.1831 84 0.7076 0.7152 0.7076 0.8412
No log 1.2113 86 0.7829 0.7125 0.7829 0.8848
No log 1.2394 88 0.8603 0.7030 0.8603 0.9275
No log 1.2676 90 0.8543 0.6957 0.8543 0.9243
No log 1.2958 92 0.8634 0.675 0.8634 0.9292
No log 1.3239 94 0.7745 0.7516 0.7745 0.8801
No log 1.3521 96 0.7257 0.7342 0.7257 0.8519
No log 1.3803 98 0.7450 0.7143 0.7450 0.8631
No log 1.4085 100 0.6965 0.7925 0.6965 0.8346
No log 1.4366 102 1.0669 0.6512 1.0669 1.0329
No log 1.4648 104 1.3621 0.5506 1.3621 1.1671
No log 1.4930 106 1.0119 0.6627 1.0119 1.0059
No log 1.5211 108 0.7427 0.7134 0.7427 0.8618
No log 1.5493 110 0.7570 0.7114 0.7570 0.8701
No log 1.5775 112 0.8250 0.7123 0.8250 0.9083
No log 1.6056 114 0.7339 0.7105 0.7339 0.8567
No log 1.6338 116 0.6620 0.7484 0.6620 0.8137
No log 1.6620 118 0.6919 0.7333 0.6919 0.8318
No log 1.6901 120 0.7974 0.6809 0.7974 0.8930
No log 1.7183 122 0.7534 0.7133 0.7534 0.8680
No log 1.7465 124 0.8260 0.6667 0.8260 0.9088
No log 1.7746 126 0.9691 0.6573 0.9691 0.9844
No log 1.8028 128 1.2706 0.6076 1.2706 1.1272
No log 1.8310 130 1.1738 0.6296 1.1738 1.0834
No log 1.8592 132 0.7725 0.7075 0.7725 0.8789
No log 1.8873 134 0.6599 0.7671 0.6599 0.8123
No log 1.9155 136 0.9657 0.6176 0.9657 0.9827
No log 1.9437 138 1.3075 0.5512 1.3075 1.1435
No log 1.9718 140 1.1479 0.5909 1.1479 1.0714
No log 2.0 142 0.7742 0.6667 0.7742 0.8799
No log 2.0282 144 0.8991 0.6667 0.8991 0.9482
No log 2.0563 146 0.8522 0.6667 0.8522 0.9232
No log 2.0845 148 0.7513 0.6897 0.7513 0.8668
No log 2.1127 150 0.6820 0.7724 0.6820 0.8258
No log 2.1408 152 0.6503 0.7808 0.6503 0.8064
No log 2.1690 154 0.6224 0.8205 0.6224 0.7889
No log 2.1972 156 0.6260 0.8121 0.6260 0.7912
No log 2.2254 158 0.6574 0.8024 0.6574 0.8108
No log 2.2535 160 0.6256 0.8193 0.6256 0.7909
No log 2.2817 162 0.6940 0.76 0.6940 0.8331
No log 2.3099 164 0.7523 0.7347 0.7523 0.8674
No log 2.3380 166 0.8291 0.7006 0.8291 0.9106
No log 2.3662 168 1.0191 0.6585 1.0191 1.0095
No log 2.3944 170 0.9327 0.6871 0.9327 0.9657
No log 2.4225 172 0.7042 0.7975 0.7042 0.8392
No log 2.4507 174 0.7423 0.7550 0.7423 0.8616
No log 2.4789 176 0.7977 0.6939 0.7977 0.8931
No log 2.5070 178 0.7738 0.7152 0.7738 0.8797
No log 2.5352 180 0.6924 0.7821 0.6924 0.8321
No log 2.5634 182 0.6789 0.8263 0.6789 0.8240
No log 2.5915 184 0.6958 0.8121 0.6958 0.8342
No log 2.6197 186 0.6921 0.7733 0.6921 0.8319
No log 2.6479 188 0.7037 0.7483 0.7037 0.8389
No log 2.6761 190 0.6854 0.7919 0.6854 0.8279
No log 2.7042 192 0.6580 0.8129 0.6580 0.8112
No log 2.7324 194 0.6326 0.8280 0.6326 0.7954
No log 2.7606 196 0.6293 0.8375 0.6293 0.7933
No log 2.7887 198 0.6149 0.8101 0.6149 0.7841
No log 2.8169 200 0.7139 0.7361 0.7139 0.8449
No log 2.8451 202 0.7376 0.7133 0.7376 0.8588
No log 2.8732 204 0.6931 0.7606 0.6931 0.8325
No log 2.9014 206 0.7521 0.7536 0.7521 0.8672
No log 2.9296 208 0.8776 0.5942 0.8776 0.9368
No log 2.9577 210 0.8764 0.6176 0.8764 0.9362
No log 2.9859 212 0.7681 0.7445 0.7681 0.8764
No log 3.0141 214 0.8025 0.7286 0.8025 0.8958
No log 3.0423 216 0.9277 0.6429 0.9277 0.9632
No log 3.0704 218 0.8339 0.6803 0.8339 0.9132
No log 3.0986 220 0.7127 0.7625 0.7127 0.8442
No log 3.1268 222 0.6401 0.7950 0.6401 0.8000
No log 3.1549 224 0.6247 0.8256 0.6247 0.7903
No log 3.1831 226 0.6302 0.8187 0.6302 0.7939
No log 3.2113 228 0.6760 0.7662 0.6760 0.8222
No log 3.2394 230 0.7648 0.7162 0.7648 0.8745
No log 3.2676 232 0.7890 0.7324 0.7890 0.8882
No log 3.2958 234 0.8307 0.7092 0.8307 0.9114
No log 3.3239 236 0.7605 0.7417 0.7605 0.8721
No log 3.3521 238 0.7184 0.7654 0.7184 0.8476
No log 3.3803 240 0.7077 0.7595 0.7077 0.8412
No log 3.4085 242 0.6715 0.7368 0.6715 0.8194
No log 3.4366 244 0.7692 0.7133 0.7692 0.8770
No log 3.4648 246 0.7637 0.7034 0.7637 0.8739
No log 3.4930 248 0.7098 0.7383 0.7098 0.8425
No log 3.5211 250 0.6425 0.7662 0.6425 0.8016
No log 3.5493 252 0.6062 0.7898 0.6062 0.7786
No log 3.5775 254 0.6108 0.7975 0.6108 0.7815
No log 3.6056 256 0.6164 0.7815 0.6164 0.7851
No log 3.6338 258 0.6228 0.7867 0.6228 0.7892
No log 3.6620 260 0.6264 0.7550 0.6264 0.7914
No log 3.6901 262 0.7816 0.7020 0.7816 0.8841
No log 3.7183 264 0.7823 0.7020 0.7823 0.8845
No log 3.7465 266 0.7141 0.7582 0.7141 0.8450
No log 3.7746 268 0.6448 0.8 0.6448 0.8030
No log 3.8028 270 0.5718 0.8182 0.5718 0.7561
No log 3.8310 272 0.6071 0.8182 0.6071 0.7792
No log 3.8592 274 0.6146 0.8302 0.6146 0.7840
No log 3.8873 276 0.5999 0.8395 0.5999 0.7745
No log 3.9155 278 0.6058 0.8395 0.6058 0.7784
No log 3.9437 280 0.6330 0.8079 0.6330 0.7956
No log 3.9718 282 0.6399 0.8383 0.6399 0.7999
No log 4.0 284 0.6765 0.8313 0.6765 0.8225
No log 4.0282 286 0.6807 0.8284 0.6807 0.8250
No log 4.0563 288 0.6664 0.8105 0.6664 0.8163
No log 4.0845 290 0.6765 0.7973 0.6765 0.8225
No log 4.1127 292 0.7003 0.7724 0.7003 0.8368
No log 4.1408 294 0.6959 0.7778 0.6959 0.8342
No log 4.1690 296 0.6735 0.7724 0.6735 0.8207
No log 4.1972 298 0.6604 0.7891 0.6604 0.8127
No log 4.2254 300 0.6599 0.7568 0.6599 0.8123
No log 4.2535 302 0.6566 0.7724 0.6566 0.8103
No log 4.2817 304 0.6749 0.7376 0.6749 0.8215
No log 4.3099 306 0.6087 0.7692 0.6087 0.7802
No log 4.3380 308 0.5563 0.7733 0.5563 0.7458
No log 4.3662 310 0.5621 0.7733 0.5621 0.7497
No log 4.3944 312 0.6031 0.7838 0.6031 0.7766
No log 4.4225 314 0.6683 0.7660 0.6683 0.8175
No log 4.4507 316 0.7202 0.7222 0.7202 0.8486
No log 4.4789 318 0.7567 0.6993 0.7567 0.8699
No log 4.5070 320 0.6714 0.7397 0.6714 0.8194
No log 4.5352 322 0.6167 0.7483 0.6167 0.7853
No log 4.5634 324 0.6021 0.7413 0.6021 0.7760
No log 4.5915 326 0.6302 0.7465 0.6302 0.7939
No log 4.6197 328 0.6671 0.7338 0.6671 0.8168
No log 4.6479 330 0.7207 0.7338 0.7207 0.8490
No log 4.6761 332 0.7850 0.6917 0.7850 0.8860
No log 4.7042 334 0.8459 0.6718 0.8459 0.9197
No log 4.7324 336 0.8375 0.6917 0.8375 0.9152
No log 4.7606 338 0.8267 0.7218 0.8267 0.9092
No log 4.7887 340 0.7516 0.7007 0.7516 0.8669
No log 4.8169 342 0.7378 0.7246 0.7378 0.8589
No log 4.8451 344 0.7503 0.7101 0.7503 0.8662
No log 4.8732 346 0.6754 0.7660 0.6754 0.8218
No log 4.9014 348 0.6048 0.7917 0.6048 0.7777
No log 4.9296 350 0.5702 0.8056 0.5702 0.7551
No log 4.9577 352 0.5725 0.7862 0.5725 0.7566
No log 4.9859 354 0.5834 0.7639 0.5834 0.7638
No log 5.0141 356 0.6025 0.7552 0.6025 0.7762
No log 5.0423 358 0.6113 0.7857 0.6113 0.7818
No log 5.0704 360 0.6178 0.7015 0.6178 0.7860
No log 5.0986 362 0.5811 0.7259 0.5811 0.7623
No log 5.1268 364 0.5376 0.7832 0.5376 0.7332
No log 5.1549 366 0.5662 0.8027 0.5662 0.7525
No log 5.1831 368 0.5988 0.7862 0.5988 0.7738
No log 5.2113 370 0.5996 0.7945 0.5996 0.7743
No log 5.2394 372 0.6177 0.7692 0.6177 0.7859
No log 5.2676 374 0.7281 0.6619 0.7281 0.8533
No log 5.2958 376 0.7164 0.7092 0.7164 0.8464
No log 5.3239 378 0.6155 0.8026 0.6155 0.7845
No log 5.3521 380 0.5868 0.8158 0.5868 0.7660
No log 5.3803 382 0.5842 0.8158 0.5842 0.7644
No log 5.4085 384 0.6170 0.7778 0.6170 0.7855
No log 5.4366 386 0.6378 0.7606 0.6378 0.7987
No log 5.4648 388 0.6164 0.7606 0.6164 0.7851
No log 5.4930 390 0.5612 0.7945 0.5612 0.7491
No log 5.5211 392 0.5138 0.8212 0.5138 0.7168
No log 5.5493 394 0.4860 0.8289 0.4860 0.6972
No log 5.5775 396 0.5209 0.8133 0.5209 0.7218
No log 5.6056 398 0.5387 0.8133 0.5387 0.7340
No log 5.6338 400 0.5625 0.8267 0.5625 0.7500
No log 5.6620 402 0.5856 0.8 0.5856 0.7653
No log 5.6901 404 0.6196 0.8079 0.6196 0.7871
No log 5.7183 406 0.5894 0.8133 0.5894 0.7678
No log 5.7465 408 0.5867 0.8079 0.5867 0.7660
No log 5.7746 410 0.5585 0.8027 0.5585 0.7474
No log 5.8028 412 0.5372 0.8 0.5372 0.7329
No log 5.8310 414 0.5506 0.8289 0.5506 0.7420
No log 5.8592 416 0.5670 0.8289 0.5670 0.7530
No log 5.8873 418 0.5902 0.8212 0.5902 0.7682
No log 5.9155 420 0.5727 0.8289 0.5727 0.7568
No log 5.9437 422 0.5496 0.8462 0.5496 0.7413
No log 5.9718 424 0.5365 0.8105 0.5365 0.7324
No log 6.0 426 0.5828 0.7550 0.5828 0.7634
No log 6.0282 428 0.5702 0.7712 0.5702 0.7551
No log 6.0563 430 0.6117 0.7632 0.6117 0.7821
No log 6.0845 432 0.6592 0.7785 0.6592 0.8119
No log 6.1127 434 0.6782 0.7671 0.6782 0.8236
No log 6.1408 436 0.7048 0.7586 0.7048 0.8395
No log 6.1690 438 0.7005 0.7671 0.7005 0.8370
No log 6.1972 440 0.6870 0.7660 0.6870 0.8288
No log 6.2254 442 0.6029 0.7746 0.6029 0.7765
No log 6.2535 444 0.5503 0.8 0.5503 0.7418
No log 6.2817 446 0.5206 0.8079 0.5206 0.7215
No log 6.3099 448 0.5255 0.8026 0.5255 0.7249
No log 6.3380 450 0.5358 0.7671 0.5358 0.7320
No log 6.3662 452 0.5886 0.7465 0.5886 0.7672
No log 6.3944 454 0.5702 0.7639 0.5702 0.7551
No log 6.4225 456 0.5574 0.7838 0.5574 0.7466
No log 6.4507 458 0.5652 0.8158 0.5652 0.7518
No log 6.4789 460 0.6189 0.8176 0.6189 0.7867
No log 6.5070 462 0.6266 0.7950 0.6266 0.7916
No log 6.5352 464 0.5642 0.8228 0.5642 0.7511
No log 6.5634 466 0.5828 0.7815 0.5828 0.7634
No log 6.5915 468 0.5950 0.7815 0.5950 0.7714
No log 6.6197 470 0.5732 0.8105 0.5732 0.7571
No log 6.6479 472 0.5958 0.8182 0.5958 0.7719
No log 6.6761 474 0.6766 0.7712 0.6766 0.8226
No log 6.7042 476 0.7478 0.7248 0.7478 0.8648
No log 6.7324 478 0.7133 0.7222 0.7133 0.8446
No log 6.7606 480 0.6301 0.7671 0.6301 0.7938
No log 6.7887 482 0.6069 0.7815 0.6069 0.7790
No log 6.8169 484 0.6234 0.7785 0.6234 0.7895
No log 6.8451 486 0.5999 0.8079 0.5999 0.7745
No log 6.8732 488 0.6161 0.8079 0.6161 0.7849
No log 6.9014 490 0.6754 0.7432 0.6754 0.8218
No log 6.9296 492 0.6818 0.7286 0.6818 0.8257
No log 6.9577 494 0.6599 0.7778 0.6599 0.8123
No log 6.9859 496 0.6683 0.7518 0.6683 0.8175
No log 7.0141 498 0.6948 0.7429 0.6948 0.8336
0.4116 7.0423 500 0.6707 0.7324 0.6707 0.8189
0.4116 7.0704 502 0.6277 0.7639 0.6277 0.7923
0.4116 7.0986 504 0.6390 0.7771 0.6390 0.7994
0.4116 7.1268 506 0.7377 0.7683 0.7377 0.8589
0.4116 7.1549 508 0.6959 0.7683 0.6959 0.8342
0.4116 7.1831 510 0.6216 0.8079 0.6216 0.7884
0.4116 7.2113 512 0.6421 0.7391 0.6421 0.8013
0.4116 7.2394 514 0.6670 0.7313 0.6670 0.8167
0.4116 7.2676 516 0.6803 0.7313 0.6803 0.8248
0.4116 7.2958 518 0.6705 0.7376 0.6705 0.8188
0.4116 7.3239 520 0.6907 0.7397 0.6907 0.8311
0.4116 7.3521 522 0.7028 0.7347 0.7028 0.8384
0.4116 7.3803 524 0.6748 0.7172 0.6748 0.8214
0.4116 7.4085 526 0.6807 0.7413 0.6807 0.8251
0.4116 7.4366 528 0.6849 0.7376 0.6849 0.8276
0.4116 7.4648 530 0.6538 0.7246 0.6538 0.8086
0.4116 7.4930 532 0.6729 0.7465 0.6729 0.8203
0.4116 7.5211 534 0.7993 0.6809 0.7993 0.8940
0.4116 7.5493 536 0.7707 0.7050 0.7707 0.8779
0.4116 7.5775 538 0.6561 0.7338 0.6561 0.8100
0.4116 7.6056 540 0.5955 0.7391 0.5955 0.7717
0.4116 7.6338 542 0.6025 0.7391 0.6025 0.7762
0.4116 7.6620 544 0.5962 0.75 0.5962 0.7722
0.4116 7.6901 546 0.6210 0.7413 0.6210 0.7880
0.4116 7.7183 548 0.6392 0.7448 0.6392 0.7995
0.4116 7.7465 550 0.6935 0.7391 0.6935 0.8328
0.4116 7.7746 552 0.7884 0.7077 0.7884 0.8879
0.4116 7.8028 554 0.7965 0.7023 0.7965 0.8925
0.4116 7.8310 556 0.7496 0.6818 0.7496 0.8658
0.4116 7.8592 558 0.6939 0.6767 0.6939 0.8330
0.4116 7.8873 560 0.6748 0.7015 0.6748 0.8214

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k15_task1_organization

Finetuned
(4019)
this model