ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k20_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4654
  • Qwk: 0.5089
  • Mse: 0.4654
  • Rmse: 0.6822

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.04 2 2.5990 -0.0262 2.5990 1.6121
No log 0.08 4 1.4044 0.0540 1.4044 1.1851
No log 0.12 6 0.9350 -0.0228 0.9350 0.9670
No log 0.16 8 0.9408 -0.0860 0.9408 0.9700
No log 0.2 10 0.8959 -0.0079 0.8959 0.9465
No log 0.24 12 0.8302 -0.0027 0.8302 0.9111
No log 0.28 14 0.8096 0.0 0.8096 0.8998
No log 0.32 16 0.8459 0.0 0.8459 0.9198
No log 0.36 18 0.8157 -0.0444 0.8157 0.9031
No log 0.4 20 0.7359 0.0717 0.7359 0.8578
No log 0.44 22 0.7123 0.2407 0.7123 0.8439
No log 0.48 24 0.6686 0.2963 0.6686 0.8177
No log 0.52 26 0.6946 0.2041 0.6946 0.8334
No log 0.56 28 0.9998 0.0975 0.9998 0.9999
No log 0.6 30 1.1047 0.1265 1.1047 1.0510
No log 0.64 32 0.9848 0.0651 0.9848 0.9924
No log 0.68 34 0.8076 0.1714 0.8076 0.8987
No log 0.72 36 0.6670 0.2412 0.6670 0.8167
No log 0.76 38 0.7052 0.3238 0.7052 0.8398
No log 0.8 40 0.7228 0.3099 0.7228 0.8502
No log 0.84 42 0.6196 0.3673 0.6196 0.7871
No log 0.88 44 0.5745 0.4709 0.5745 0.7579
No log 0.92 46 0.5651 0.4538 0.5651 0.7517
No log 0.96 48 0.5628 0.5470 0.5628 0.7502
No log 1.0 50 0.7374 0.3693 0.7374 0.8587
No log 1.04 52 1.0917 0.2214 1.0917 1.0448
No log 1.08 54 1.1288 0.2439 1.1288 1.0625
No log 1.12 56 0.9953 0.2658 0.9953 0.9977
No log 1.16 58 0.9210 0.2273 0.9210 0.9597
No log 1.2 60 0.8072 0.2817 0.8072 0.8985
No log 1.24 62 0.6368 0.3829 0.6368 0.7980
No log 1.28 64 0.6022 0.3633 0.6022 0.7760
No log 1.32 66 0.5991 0.2641 0.5991 0.7740
No log 1.3600 68 0.5737 0.4007 0.5737 0.7574
No log 1.4 70 0.5645 0.4238 0.5645 0.7513
No log 1.44 72 0.5500 0.3274 0.5500 0.7416
No log 1.48 74 0.5848 0.2817 0.5848 0.7647
No log 1.52 76 0.6111 0.2851 0.6111 0.7817
No log 1.56 78 0.6344 0.3506 0.6344 0.7965
No log 1.6 80 0.6331 0.3477 0.6331 0.7957
No log 1.6400 82 0.6274 0.3446 0.6274 0.7921
No log 1.6800 84 0.6045 0.2890 0.6045 0.7775
No log 1.72 86 0.6001 0.2890 0.6001 0.7747
No log 1.76 88 0.6040 0.3866 0.6040 0.7772
No log 1.8 90 0.6018 0.3701 0.6018 0.7758
No log 1.8400 92 0.5700 0.4729 0.5700 0.7550
No log 1.88 94 0.5897 0.4677 0.5897 0.7679
No log 1.92 96 0.6277 0.4537 0.6277 0.7923
No log 1.96 98 0.6177 0.4969 0.6177 0.7859
No log 2.0 100 0.5121 0.4358 0.5121 0.7156
No log 2.04 102 0.5277 0.4855 0.5277 0.7264
No log 2.08 104 0.5263 0.4769 0.5263 0.7255
No log 2.12 106 0.5053 0.4973 0.5053 0.7108
No log 2.16 108 0.6179 0.5267 0.6179 0.7861
No log 2.2 110 0.8142 0.5167 0.8142 0.9023
No log 2.24 112 0.8153 0.4993 0.8153 0.9029
No log 2.2800 114 0.8404 0.3942 0.8404 0.9167
No log 2.32 116 0.7278 0.4430 0.7278 0.8531
No log 2.36 118 0.5734 0.5763 0.5734 0.7573
No log 2.4 120 0.4917 0.5022 0.4917 0.7012
No log 2.44 122 0.6217 0.4648 0.6217 0.7885
No log 2.48 124 0.6749 0.4716 0.6749 0.8215
No log 2.52 126 0.6093 0.4568 0.6093 0.7806
No log 2.56 128 0.5706 0.5050 0.5706 0.7554
No log 2.6 130 0.5423 0.4608 0.5423 0.7364
No log 2.64 132 0.5573 0.5117 0.5573 0.7465
No log 2.68 134 0.6152 0.6154 0.6152 0.7844
No log 2.7200 136 0.5795 0.6154 0.5795 0.7612
No log 2.76 138 0.5338 0.5589 0.5338 0.7306
No log 2.8 140 0.6603 0.5614 0.6603 0.8126
No log 2.84 142 0.7495 0.4199 0.7495 0.8657
No log 2.88 144 0.6996 0.3976 0.6996 0.8364
No log 2.92 146 0.5926 0.4051 0.5926 0.7698
No log 2.96 148 0.5438 0.5463 0.5438 0.7374
No log 3.0 150 0.5440 0.4762 0.5440 0.7375
No log 3.04 152 0.5880 0.3890 0.5880 0.7668
No log 3.08 154 0.7044 0.3475 0.7044 0.8393
No log 3.12 156 0.7134 0.3475 0.7134 0.8446
No log 3.16 158 0.6073 0.3914 0.6073 0.7793
No log 3.2 160 0.5393 0.5357 0.5393 0.7344
No log 3.24 162 0.5572 0.5855 0.5572 0.7465
No log 3.2800 164 0.5962 0.5765 0.5962 0.7722
No log 3.32 166 0.6744 0.5441 0.6744 0.8212
No log 3.36 168 0.7142 0.5441 0.7142 0.8451
No log 3.4 170 0.7065 0.5441 0.7065 0.8406
No log 3.44 172 0.5857 0.6195 0.5857 0.7653
No log 3.48 174 0.5410 0.4924 0.5410 0.7355
No log 3.52 176 0.5383 0.5067 0.5383 0.7337
No log 3.56 178 0.5493 0.6020 0.5493 0.7412
No log 3.6 180 0.6201 0.5500 0.6201 0.7874
No log 3.64 182 0.6037 0.5298 0.6037 0.7770
No log 3.68 184 0.5255 0.5826 0.5255 0.7249
No log 3.7200 186 0.5462 0.5157 0.5462 0.7391
No log 3.76 188 0.6825 0.3824 0.6825 0.8261
No log 3.8 190 0.7076 0.3645 0.7076 0.8412
No log 3.84 192 0.6400 0.4453 0.6400 0.8000
No log 3.88 194 0.5672 0.4557 0.5672 0.7532
No log 3.92 196 0.5645 0.4974 0.5645 0.7513
No log 3.96 198 0.5693 0.6041 0.5693 0.7545
No log 4.0 200 0.5671 0.6041 0.5671 0.7530
No log 4.04 202 0.5552 0.5377 0.5552 0.7451
No log 4.08 204 0.5846 0.4302 0.5846 0.7646
No log 4.12 206 0.5835 0.4302 0.5835 0.7639
No log 4.16 208 0.5579 0.4374 0.5579 0.7469
No log 4.2 210 0.5432 0.5174 0.5432 0.7370
No log 4.24 212 0.6097 0.5966 0.6097 0.7808
No log 4.28 214 0.6030 0.5780 0.6030 0.7766
No log 4.32 216 0.5629 0.5547 0.5629 0.7502
No log 4.36 218 0.5808 0.5780 0.5808 0.7621
No log 4.4 220 0.5263 0.5939 0.5263 0.7255
No log 4.44 222 0.5268 0.5505 0.5268 0.7258
No log 4.48 224 0.5706 0.6030 0.5706 0.7554
No log 4.52 226 0.5730 0.6207 0.5730 0.7570
No log 4.5600 228 0.5450 0.5756 0.5450 0.7383
No log 4.6 230 0.5358 0.5552 0.5358 0.7320
No log 4.64 232 0.5484 0.5353 0.5484 0.7405
No log 4.68 234 0.5220 0.4907 0.5220 0.7225
No log 4.72 236 0.5025 0.5993 0.5025 0.7089
No log 4.76 238 0.4956 0.5993 0.4956 0.7040
No log 4.8 240 0.5025 0.5050 0.5025 0.7089
No log 4.84 242 0.5616 0.5275 0.5616 0.7494
No log 4.88 244 0.5719 0.4862 0.5719 0.7562
No log 4.92 246 0.5145 0.5142 0.5145 0.7173
No log 4.96 248 0.5314 0.6537 0.5314 0.7290
No log 5.0 250 0.5343 0.6036 0.5343 0.7310
No log 5.04 252 0.4942 0.6407 0.4942 0.7030
No log 5.08 254 0.5009 0.4828 0.5009 0.7077
No log 5.12 256 0.4997 0.4828 0.4997 0.7069
No log 5.16 258 0.5053 0.5272 0.5053 0.7109
No log 5.2 260 0.4720 0.5177 0.4720 0.6870
No log 5.24 262 0.4700 0.6197 0.4700 0.6856
No log 5.28 264 0.4561 0.5617 0.4561 0.6754
No log 5.32 266 0.4486 0.5208 0.4486 0.6698
No log 5.36 268 0.4484 0.5286 0.4484 0.6696
No log 5.4 270 0.4602 0.6168 0.4602 0.6784
No log 5.44 272 0.4554 0.6068 0.4554 0.6748
No log 5.48 274 0.4498 0.5195 0.4498 0.6707
No log 5.52 276 0.4498 0.5580 0.4498 0.6706
No log 5.5600 278 0.4478 0.5970 0.4478 0.6692
No log 5.6 280 0.4934 0.5455 0.4934 0.7025
No log 5.64 282 0.5884 0.4880 0.5884 0.7671
No log 5.68 284 0.5381 0.5149 0.5381 0.7335
No log 5.72 286 0.4748 0.5985 0.4748 0.6891
No log 5.76 288 0.5731 0.5664 0.5731 0.7570
No log 5.8 290 0.6126 0.6230 0.6126 0.7827
No log 5.84 292 0.5258 0.6379 0.5258 0.7251
No log 5.88 294 0.4694 0.5753 0.4694 0.6851
No log 5.92 296 0.4737 0.6108 0.4737 0.6883
No log 5.96 298 0.4664 0.6014 0.4664 0.6830
No log 6.0 300 0.4574 0.5915 0.4574 0.6763
No log 6.04 302 0.4475 0.5915 0.4475 0.6690
No log 6.08 304 0.4410 0.6407 0.4410 0.6641
No log 6.12 306 0.4369 0.6146 0.4369 0.6610
No log 6.16 308 0.4637 0.6720 0.4637 0.6810
No log 6.2 310 0.5400 0.5338 0.5400 0.7348
No log 6.24 312 0.5346 0.5639 0.5346 0.7311
No log 6.28 314 0.5184 0.5338 0.5184 0.7200
No log 6.32 316 0.4669 0.6052 0.4669 0.6833
No log 6.36 318 0.4345 0.6040 0.4345 0.6592
No log 6.4 320 0.4685 0.5765 0.4685 0.6844
No log 6.44 322 0.4687 0.5836 0.4687 0.6846
No log 6.48 324 0.4241 0.6007 0.4241 0.6512
No log 6.52 326 0.4249 0.6242 0.4249 0.6518
No log 6.5600 328 0.4177 0.5875 0.4177 0.6463
No log 6.6 330 0.4374 0.5796 0.4374 0.6614
No log 6.64 332 0.4926 0.5808 0.4926 0.7019
No log 6.68 334 0.5374 0.5543 0.5374 0.7331
No log 6.72 336 0.4996 0.5466 0.4996 0.7068
No log 6.76 338 0.4387 0.4817 0.4387 0.6623
No log 6.8 340 0.4367 0.5223 0.4367 0.6608
No log 6.84 342 0.4351 0.4968 0.4351 0.6596
No log 6.88 344 0.4662 0.5512 0.4662 0.6828
No log 6.92 346 0.5617 0.5900 0.5617 0.7495
No log 6.96 348 0.5558 0.5778 0.5558 0.7455
No log 7.0 350 0.5339 0.5874 0.5339 0.7307
No log 7.04 352 0.4959 0.5553 0.4959 0.7042
No log 7.08 354 0.4735 0.5841 0.4735 0.6881
No log 7.12 356 0.4666 0.5321 0.4666 0.6831
No log 7.16 358 0.4735 0.5587 0.4735 0.6881
No log 7.2 360 0.4661 0.5356 0.4661 0.6827
No log 7.24 362 0.4604 0.5319 0.4604 0.6786
No log 7.28 364 0.4683 0.5765 0.4683 0.6843
No log 7.32 366 0.4941 0.5721 0.4941 0.7029
No log 7.36 368 0.4733 0.5692 0.4733 0.6880
No log 7.4 370 0.4650 0.5452 0.4650 0.6819
No log 7.44 372 0.4655 0.5421 0.4655 0.6823
No log 7.48 374 0.4640 0.5492 0.4640 0.6812
No log 7.52 376 0.4687 0.5087 0.4687 0.6846
No log 7.5600 378 0.4686 0.5269 0.4686 0.6845
No log 7.6 380 0.5036 0.5223 0.5036 0.7096
No log 7.64 382 0.5424 0.5291 0.5424 0.7365
No log 7.68 384 0.5172 0.5291 0.5172 0.7192
No log 7.72 386 0.4882 0.5306 0.4882 0.6987
No log 7.76 388 0.4848 0.5596 0.4848 0.6963
No log 7.8 390 0.4865 0.5846 0.4865 0.6975
No log 7.84 392 0.5009 0.5067 0.5009 0.7077
No log 7.88 394 0.5455 0.5431 0.5455 0.7386
No log 7.92 396 0.5367 0.5553 0.5367 0.7326
No log 7.96 398 0.4911 0.5167 0.4911 0.7008
No log 8.0 400 0.4963 0.5779 0.4963 0.7045
No log 8.04 402 0.5152 0.5313 0.5152 0.7178
No log 8.08 404 0.4899 0.5908 0.4899 0.6999
No log 8.12 406 0.4704 0.5915 0.4704 0.6858
No log 8.16 408 0.5126 0.5692 0.5126 0.7160
No log 8.2 410 0.5238 0.5692 0.5238 0.7237
No log 8.24 412 0.4932 0.5753 0.4932 0.7023
No log 8.28 414 0.4875 0.5753 0.4875 0.6982
No log 8.32 416 0.4847 0.5768 0.4847 0.6962
No log 8.36 418 0.5042 0.4901 0.5042 0.7101
No log 8.4 420 0.5215 0.4704 0.5215 0.7221
No log 8.44 422 0.5364 0.4622 0.5364 0.7324
No log 8.48 424 0.4975 0.4743 0.4975 0.7053
No log 8.52 426 0.4668 0.6255 0.4668 0.6832
No log 8.56 428 0.4660 0.5826 0.4660 0.6826
No log 8.6 430 0.4536 0.5985 0.4536 0.6735
No log 8.64 432 0.4792 0.5404 0.4792 0.6922
No log 8.68 434 0.5435 0.5179 0.5435 0.7372
No log 8.72 436 0.5753 0.5862 0.5753 0.7585
No log 8.76 438 0.5111 0.5886 0.5111 0.7149
No log 8.8 440 0.4567 0.6423 0.4567 0.6758
No log 8.84 442 0.4751 0.5983 0.4751 0.6892
No log 8.88 444 0.5024 0.5922 0.5024 0.7088
No log 8.92 446 0.4647 0.5733 0.4647 0.6817
No log 8.96 448 0.4520 0.5811 0.4520 0.6723
No log 9.0 450 0.4545 0.5343 0.4545 0.6741
No log 9.04 452 0.4583 0.4817 0.4583 0.6770
No log 9.08 454 0.4510 0.5510 0.4510 0.6716
No log 9.12 456 0.4847 0.5455 0.4847 0.6962
No log 9.16 458 0.5792 0.5557 0.5792 0.7610
No log 9.2 460 0.6408 0.5368 0.6408 0.8005
No log 9.24 462 0.5718 0.5736 0.5718 0.7562
No log 9.28 464 0.4727 0.5455 0.4727 0.6876
No log 9.32 466 0.4473 0.5085 0.4473 0.6688
No log 9.36 468 0.4857 0.4803 0.4857 0.6969
No log 9.4 470 0.4885 0.4908 0.4885 0.6990
No log 9.44 472 0.4673 0.5719 0.4673 0.6836
No log 9.48 474 0.4575 0.5352 0.4575 0.6764
No log 9.52 476 0.4985 0.5849 0.4985 0.7060
No log 9.56 478 0.5350 0.5970 0.5350 0.7314
No log 9.6 480 0.5066 0.6060 0.5066 0.7118
No log 9.64 482 0.4921 0.5898 0.4921 0.7015
No log 9.68 484 0.4763 0.5932 0.4763 0.6902
No log 9.72 486 0.4715 0.5779 0.4715 0.6867
No log 9.76 488 0.4690 0.5861 0.4690 0.6848
No log 9.8 490 0.4876 0.5340 0.4876 0.6983
No log 9.84 492 0.4910 0.5208 0.4910 0.7007
No log 9.88 494 0.4840 0.5227 0.4840 0.6957
No log 9.92 496 0.4748 0.5414 0.4748 0.6891
No log 9.96 498 0.4703 0.5414 0.4703 0.6858
0.3902 10.0 500 0.4652 0.5633 0.4652 0.6821
0.3902 10.04 502 0.4955 0.4986 0.4955 0.7039
0.3902 10.08 504 0.5420 0.5166 0.5420 0.7362
0.3902 10.12 506 0.5402 0.5166 0.5402 0.7350
0.3902 10.16 508 0.4837 0.4937 0.4837 0.6955
0.3902 10.2 510 0.4659 0.5868 0.4659 0.6825
0.3902 10.24 512 0.4723 0.5642 0.4723 0.6872
0.3902 10.28 514 0.4684 0.5877 0.4684 0.6844
0.3902 10.32 516 0.4669 0.4656 0.4669 0.6833
0.3902 10.36 518 0.4654 0.4656 0.4654 0.6822
0.3902 10.4 520 0.4673 0.4656 0.4673 0.6836
0.3902 10.44 522 0.4613 0.5042 0.4613 0.6792
0.3902 10.48 524 0.4654 0.5089 0.4654 0.6822

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k20_task7_organization

Finetuned
(4023)
this model