ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k11_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4477
  • Qwk: 0.5882
  • Mse: 0.4477
  • Rmse: 0.6691

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 2.4713 -0.0262 2.4713 1.5720
No log 0.1379 4 1.3112 0.0998 1.3112 1.1451
No log 0.2069 6 0.8126 0.0535 0.8126 0.9014
No log 0.2759 8 0.9384 0.0927 0.9384 0.9687
No log 0.3448 10 0.9361 0.2939 0.9361 0.9675
No log 0.4138 12 0.9504 0.2601 0.9504 0.9749
No log 0.4828 14 0.8084 0.2285 0.8084 0.8991
No log 0.5517 16 0.7364 -0.0027 0.7364 0.8582
No log 0.6207 18 0.7324 0.0898 0.7324 0.8558
No log 0.6897 20 0.7246 0.2494 0.7246 0.8513
No log 0.7586 22 0.6418 0.1277 0.6418 0.8011
No log 0.8276 24 0.6349 0.1604 0.6349 0.7968
No log 0.8966 26 0.6258 0.2002 0.6258 0.7911
No log 0.9655 28 0.6073 0.2783 0.6073 0.7793
No log 1.0345 30 0.6091 0.3426 0.6091 0.7804
No log 1.1034 32 0.5988 0.3377 0.5988 0.7738
No log 1.1724 34 0.7307 0.2886 0.7307 0.8548
No log 1.2414 36 0.8048 0.3094 0.8048 0.8971
No log 1.3103 38 0.7675 0.2817 0.7675 0.8761
No log 1.3793 40 0.6645 0.3999 0.6645 0.8152
No log 1.4483 42 0.5600 0.5361 0.5600 0.7483
No log 1.5172 44 0.6354 0.1935 0.6354 0.7971
No log 1.5862 46 0.8653 0.0717 0.8653 0.9302
No log 1.6552 48 0.8344 0.1268 0.8344 0.9134
No log 1.7241 50 0.6252 0.2995 0.6252 0.7907
No log 1.7931 52 0.5520 0.4938 0.5520 0.7430
No log 1.8621 54 0.6968 0.4280 0.6968 0.8347
No log 1.9310 56 0.9316 0.2009 0.9316 0.9652
No log 2.0 58 0.7644 0.3484 0.7644 0.8743
No log 2.0690 60 0.5697 0.4229 0.5697 0.7548
No log 2.1379 62 0.7542 0.3606 0.7542 0.8685
No log 2.2069 64 0.7940 0.3473 0.7940 0.8911
No log 2.2759 66 0.6665 0.2027 0.6665 0.8164
No log 2.3448 68 0.6484 0.2892 0.6484 0.8053
No log 2.4138 70 0.6588 0.2892 0.6588 0.8117
No log 2.4828 72 0.6602 0.2374 0.6602 0.8125
No log 2.5517 74 0.6709 0.2506 0.6709 0.8191
No log 2.6207 76 0.7578 0.2692 0.7578 0.8705
No log 2.6897 78 0.8454 0.3042 0.8454 0.9195
No log 2.7586 80 0.7795 0.3302 0.7795 0.8829
No log 2.8276 82 0.7342 0.2995 0.7342 0.8569
No log 2.8966 84 0.6643 0.3947 0.6643 0.8151
No log 2.9655 86 0.6658 0.4448 0.6658 0.8160
No log 3.0345 88 0.6493 0.4019 0.6493 0.8058
No log 3.1034 90 0.7082 0.4134 0.7082 0.8416
No log 3.1724 92 0.6085 0.5345 0.6085 0.7801
No log 3.2414 94 0.5784 0.4898 0.5784 0.7605
No log 3.3103 96 0.6340 0.4698 0.6340 0.7963
No log 3.3793 98 0.5717 0.4494 0.5717 0.7561
No log 3.4483 100 0.6084 0.4905 0.6084 0.7800
No log 3.5172 102 0.7665 0.4462 0.7665 0.8755
No log 3.5862 104 0.7339 0.4462 0.7339 0.8567
No log 3.6552 106 0.5707 0.5067 0.5707 0.7555
No log 3.7241 108 0.5439 0.4918 0.5439 0.7375
No log 3.7931 110 0.6868 0.5146 0.6868 0.8287
No log 3.8621 112 0.7512 0.4784 0.7512 0.8667
No log 3.9310 114 0.6646 0.5704 0.6646 0.8152
No log 4.0 116 0.6148 0.5285 0.6148 0.7841
No log 4.0690 118 0.5934 0.5397 0.5934 0.7703
No log 4.1379 120 0.5865 0.5646 0.5865 0.7658
No log 4.2069 122 0.5621 0.4753 0.5621 0.7497
No log 4.2759 124 0.5439 0.5617 0.5439 0.7375
No log 4.3448 126 0.5358 0.5184 0.5358 0.7320
No log 4.4138 128 0.5188 0.4478 0.5188 0.7203
No log 4.4828 130 0.5472 0.5587 0.5472 0.7398
No log 4.5517 132 0.5155 0.4762 0.5155 0.7180
No log 4.6207 134 0.5184 0.4828 0.5184 0.7200
No log 4.6897 136 0.5290 0.4828 0.5290 0.7273
No log 4.7586 138 0.5448 0.5157 0.5448 0.7381
No log 4.8276 140 0.5847 0.5195 0.5847 0.7647
No log 4.8966 142 0.5500 0.5157 0.5500 0.7416
No log 4.9655 144 0.5288 0.5184 0.5288 0.7272
No log 5.0345 146 0.5227 0.5413 0.5227 0.7230
No log 5.1034 148 0.5281 0.4596 0.5281 0.7267
No log 5.1724 150 0.5248 0.5798 0.5248 0.7244
No log 5.2414 152 0.5274 0.5714 0.5274 0.7262
No log 5.3103 154 0.5338 0.5475 0.5338 0.7306
No log 5.3793 156 0.5509 0.4147 0.5509 0.7423
No log 5.4483 158 0.6435 0.4243 0.6435 0.8022
No log 5.5172 160 0.6564 0.4243 0.6564 0.8102
No log 5.5862 162 0.5670 0.4432 0.5670 0.7530
No log 5.6552 164 0.5724 0.5286 0.5724 0.7566
No log 5.7241 166 0.5859 0.5697 0.5859 0.7655
No log 5.7931 168 0.6038 0.5190 0.6038 0.7771
No log 5.8621 170 0.6303 0.5220 0.6303 0.7939
No log 5.9310 172 0.6049 0.5220 0.6049 0.7778
No log 6.0 174 0.5347 0.5596 0.5347 0.7312
No log 6.0690 176 0.5282 0.5413 0.5282 0.7268
No log 6.1379 178 0.5217 0.5672 0.5217 0.7223
No log 6.2069 180 0.5375 0.5205 0.5375 0.7332
No log 6.2759 182 0.5952 0.5244 0.5952 0.7715
No log 6.3448 184 0.5982 0.5273 0.5982 0.7734
No log 6.4138 186 0.6063 0.5483 0.6063 0.7787
No log 6.4828 188 0.5968 0.5483 0.5968 0.7725
No log 6.5517 190 0.5583 0.5143 0.5583 0.7472
No log 6.6207 192 0.7077 0.3973 0.7077 0.8413
No log 6.6897 194 1.0660 0.3455 1.0660 1.0325
No log 6.7586 196 1.0889 0.2601 1.0889 1.0435
No log 6.8276 198 0.8660 0.3269 0.8660 0.9306
No log 6.8966 200 0.6400 0.3399 0.6400 0.8000
No log 6.9655 202 0.5973 0.4345 0.5973 0.7728
No log 7.0345 204 0.6270 0.4821 0.6270 0.7918
No log 7.1034 206 0.5828 0.4849 0.5828 0.7634
No log 7.1724 208 0.5997 0.3665 0.5997 0.7744
No log 7.2414 210 0.6342 0.3936 0.6342 0.7963
No log 7.3103 212 0.6472 0.4335 0.6472 0.8045
No log 7.3793 214 0.5916 0.4186 0.5916 0.7692
No log 7.4483 216 0.5353 0.4591 0.5353 0.7317
No log 7.5172 218 0.5467 0.4821 0.5467 0.7394
No log 7.5862 220 0.5656 0.5065 0.5656 0.7521
No log 7.6552 222 0.5211 0.5796 0.5211 0.7219
No log 7.7241 224 0.5079 0.5596 0.5079 0.7127
No log 7.7931 226 0.5057 0.5953 0.5057 0.7111
No log 7.8621 228 0.5154 0.5796 0.5154 0.7179
No log 7.9310 230 0.5107 0.5796 0.5107 0.7146
No log 8.0 232 0.4926 0.5877 0.4926 0.7019
No log 8.0690 234 0.4859 0.5985 0.4859 0.6971
No log 8.1379 236 0.4888 0.5904 0.4888 0.6992
No log 8.2069 238 0.4964 0.6465 0.4964 0.7046
No log 8.2759 240 0.5124 0.6361 0.5124 0.7158
No log 8.3448 242 0.5104 0.6130 0.5104 0.7144
No log 8.4138 244 0.4939 0.6020 0.4939 0.7028
No log 8.4828 246 0.4945 0.5283 0.4945 0.7032
No log 8.5517 248 0.4839 0.5877 0.4839 0.6956
No log 8.6207 250 0.4803 0.5556 0.4803 0.6931
No log 8.6897 252 0.4878 0.5986 0.4878 0.6984
No log 8.7586 254 0.5057 0.5715 0.5057 0.7111
No log 8.8276 256 0.4907 0.6082 0.4907 0.7005
No log 8.8966 258 0.4827 0.5714 0.4827 0.6947
No log 8.9655 260 0.4870 0.5265 0.4870 0.6978
No log 9.0345 262 0.4969 0.5432 0.4969 0.7049
No log 9.1034 264 0.4916 0.5463 0.4916 0.7011
No log 9.1724 266 0.4940 0.6142 0.4940 0.7029
No log 9.2414 268 0.5125 0.5708 0.5125 0.7159
No log 9.3103 270 0.5330 0.5166 0.5330 0.7301
No log 9.3793 272 0.5099 0.5897 0.5099 0.7141
No log 9.4483 274 0.5129 0.5345 0.5129 0.7162
No log 9.5172 276 0.5037 0.5123 0.5037 0.7097
No log 9.5862 278 0.4825 0.6555 0.4825 0.6947
No log 9.6552 280 0.4839 0.6154 0.4839 0.6957
No log 9.7241 282 0.4843 0.5861 0.4843 0.6959
No log 9.7931 284 0.5034 0.5587 0.5034 0.7095
No log 9.8621 286 0.5102 0.5666 0.5102 0.7143
No log 9.9310 288 0.5439 0.4931 0.5439 0.7375
No log 10.0 290 0.5163 0.5438 0.5163 0.7186
No log 10.0690 292 0.5246 0.5438 0.5246 0.7243
No log 10.1379 294 0.5295 0.4931 0.5295 0.7277
No log 10.2069 296 0.5055 0.5587 0.5055 0.7110
No log 10.2759 298 0.4876 0.5510 0.4876 0.6983
No log 10.3448 300 0.4749 0.5798 0.4749 0.6891
No log 10.4138 302 0.4717 0.5798 0.4717 0.6868
No log 10.4828 304 0.4677 0.6017 0.4677 0.6839
No log 10.5517 306 0.4763 0.5912 0.4763 0.6901
No log 10.6207 308 0.4620 0.6854 0.4620 0.6797
No log 10.6897 310 0.4656 0.5904 0.4656 0.6824
No log 10.7586 312 0.4661 0.5589 0.4661 0.6827
No log 10.8276 314 0.5060 0.5411 0.5060 0.7114
No log 10.8966 316 0.5065 0.5254 0.5065 0.7117
No log 10.9655 318 0.4874 0.5723 0.4874 0.6981
No log 11.0345 320 0.4658 0.5589 0.4658 0.6825
No log 11.1034 322 0.4496 0.6105 0.4496 0.6705
No log 11.1724 324 0.4654 0.6595 0.4654 0.6822
No log 11.2414 326 0.4634 0.6772 0.4634 0.6808
No log 11.3103 328 0.4345 0.6184 0.4345 0.6592
No log 11.3793 330 0.4749 0.5237 0.4749 0.6891
No log 11.4483 332 0.5150 0.5067 0.5150 0.7177
No log 11.5172 334 0.4893 0.5017 0.4893 0.6995
No log 11.5862 336 0.4430 0.6200 0.4430 0.6656
No log 11.6552 338 0.4400 0.6229 0.4400 0.6633
No log 11.7241 340 0.4496 0.6489 0.4496 0.6705
No log 11.7931 342 0.4307 0.6577 0.4307 0.6563
No log 11.8621 344 0.4234 0.6010 0.4234 0.6507
No log 11.9310 346 0.4569 0.5801 0.4569 0.6760
No log 12.0 348 0.4397 0.6065 0.4397 0.6631
No log 12.0690 350 0.4514 0.6210 0.4514 0.6719
No log 12.1379 352 0.5128 0.5779 0.5128 0.7161
No log 12.2069 354 0.5045 0.5855 0.5045 0.7103
No log 12.2759 356 0.4599 0.5861 0.4599 0.6781
No log 12.3448 358 0.4970 0.5252 0.4970 0.7050
No log 12.4138 360 0.5993 0.4815 0.5993 0.7741
No log 12.4828 362 0.6608 0.4726 0.6608 0.8129
No log 12.5517 364 0.5931 0.5233 0.5931 0.7701
No log 12.6207 366 0.4987 0.4753 0.4987 0.7062
No log 12.6897 368 0.4743 0.4147 0.4743 0.6887
No log 12.7586 370 0.4750 0.5133 0.4750 0.6892
No log 12.8276 372 0.4718 0.4768 0.4718 0.6868
No log 12.8966 374 0.4952 0.5233 0.4952 0.7037
No log 12.9655 376 0.5749 0.4728 0.5749 0.7582
No log 13.0345 378 0.6155 0.4644 0.6155 0.7845
No log 13.1034 380 0.5617 0.4728 0.5617 0.7495
No log 13.1724 382 0.4800 0.5177 0.4800 0.6928
No log 13.2414 384 0.4656 0.5868 0.4656 0.6823
No log 13.3103 386 0.4676 0.6293 0.4676 0.6838
No log 13.3793 388 0.4793 0.5289 0.4793 0.6923
No log 13.4483 390 0.5125 0.5177 0.5125 0.7159
No log 13.5172 392 0.5131 0.5177 0.5131 0.7163
No log 13.5862 394 0.4683 0.5768 0.4683 0.6843
No log 13.6552 396 0.4780 0.5781 0.4780 0.6914
No log 13.7241 398 0.5475 0.5315 0.5475 0.7399
No log 13.7931 400 0.5479 0.5230 0.5479 0.7402
No log 13.8621 402 0.5044 0.5945 0.5044 0.7102
No log 13.9310 404 0.4800 0.5869 0.4800 0.6928
No log 14.0 406 0.4838 0.5765 0.4838 0.6956
No log 14.0690 408 0.4838 0.5605 0.4838 0.6956
No log 14.1379 410 0.4798 0.5413 0.4798 0.6927
No log 14.2069 412 0.4710 0.5267 0.4710 0.6863
No log 14.2759 414 0.4636 0.5344 0.4636 0.6809
No log 14.3448 416 0.4624 0.4878 0.4624 0.6800
No log 14.4138 418 0.4612 0.5522 0.4612 0.6791
No log 14.4828 420 0.4626 0.5430 0.4626 0.6801
No log 14.5517 422 0.4643 0.5522 0.4643 0.6814
No log 14.6207 424 0.4661 0.5549 0.4661 0.6827
No log 14.6897 426 0.4686 0.5488 0.4686 0.6845
No log 14.7586 428 0.4796 0.5457 0.4796 0.6925
No log 14.8276 430 0.4839 0.5512 0.4839 0.6956
No log 14.8966 432 0.5060 0.5219 0.5060 0.7113
No log 14.9655 434 0.5378 0.5233 0.5378 0.7334
No log 15.0345 436 0.5062 0.5091 0.5062 0.7115
No log 15.1034 438 0.4856 0.4685 0.4856 0.6968
No log 15.1724 440 0.4624 0.6307 0.4624 0.6800
No log 15.2414 442 0.4516 0.6435 0.4516 0.6720
No log 15.3103 444 0.4481 0.6228 0.4481 0.6694
No log 15.3793 446 0.4722 0.5752 0.4722 0.6872
No log 15.4483 448 0.4842 0.5468 0.4842 0.6959
No log 15.5172 450 0.4541 0.6228 0.4541 0.6739
No log 15.5862 452 0.4416 0.5979 0.4416 0.6646
No log 15.6552 454 0.4489 0.6950 0.4489 0.6700
No log 15.7241 456 0.4514 0.6771 0.4514 0.6719
No log 15.7931 458 0.4427 0.5979 0.4427 0.6654
No log 15.8621 460 0.4465 0.6142 0.4465 0.6682
No log 15.9310 462 0.4492 0.6142 0.4492 0.6702
No log 16.0 464 0.4639 0.6201 0.4639 0.6811
No log 16.0690 466 0.4610 0.5698 0.4610 0.6790
No log 16.1379 468 0.4508 0.6171 0.4508 0.6714
No log 16.2069 470 0.4398 0.6096 0.4398 0.6632
No log 16.2759 472 0.4412 0.6010 0.4412 0.6642
No log 16.3448 474 0.4472 0.6303 0.4472 0.6687
No log 16.4138 476 0.4661 0.5349 0.4661 0.6827
No log 16.4828 478 0.4894 0.4835 0.4894 0.6996
No log 16.5517 480 0.5003 0.4835 0.5003 0.7073
No log 16.6207 482 0.4742 0.4925 0.4742 0.6886
No log 16.6897 484 0.4523 0.5815 0.4523 0.6726
No log 16.7586 486 0.4738 0.5666 0.4738 0.6883
No log 16.8276 488 0.4748 0.5747 0.4748 0.6891
No log 16.8966 490 0.4611 0.5574 0.4611 0.6791
No log 16.9655 492 0.4675 0.5209 0.4675 0.6837
No log 17.0345 494 0.4822 0.5252 0.4822 0.6944
No log 17.1034 496 0.4773 0.5252 0.4773 0.6909
No log 17.1724 498 0.4590 0.5420 0.4590 0.6775
0.3438 17.2414 500 0.4402 0.6339 0.4402 0.6635
0.3438 17.3103 502 0.4441 0.5596 0.4441 0.6664
0.3438 17.3793 504 0.4442 0.6007 0.4442 0.6665
0.3438 17.4483 506 0.4452 0.6007 0.4452 0.6672
0.3438 17.5172 508 0.4364 0.6105 0.4364 0.6606
0.3438 17.5862 510 0.4477 0.5882 0.4477 0.6691

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k11_task7_organization

Finetuned
(4023)
this model