ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k10_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5473
  • Qwk: 0.4300
  • Mse: 0.5473
  • Rmse: 0.7398

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0741 2 2.5262 -0.0262 2.5262 1.5894
No log 0.1481 4 1.3142 0.1262 1.3142 1.1464
No log 0.2222 6 1.0509 -0.0970 1.0509 1.0251
No log 0.2963 8 0.8892 -0.0764 0.8892 0.9430
No log 0.3704 10 0.8148 0.0026 0.8148 0.9027
No log 0.4444 12 0.8081 0.0053 0.8081 0.8989
No log 0.5185 14 0.8261 0.0481 0.8261 0.9089
No log 0.5926 16 0.8230 0.0481 0.8230 0.9072
No log 0.6667 18 0.8115 0.0 0.8115 0.9009
No log 0.7407 20 0.7755 0.0 0.7755 0.8806
No log 0.8148 22 0.7757 0.0 0.7757 0.8808
No log 0.8889 24 0.8157 0.0393 0.8157 0.9032
No log 0.9630 26 0.8129 -0.0027 0.8129 0.9016
No log 1.0370 28 0.7998 0.0 0.7998 0.8943
No log 1.1111 30 0.8472 -0.0027 0.8472 0.9205
No log 1.1852 32 0.7863 0.0 0.7863 0.8867
No log 1.2593 34 0.7741 0.0937 0.7741 0.8798
No log 1.3333 36 0.7823 0.0937 0.7823 0.8845
No log 1.4074 38 0.7801 0.0481 0.7801 0.8832
No log 1.4815 40 0.7897 0.0937 0.7897 0.8887
No log 1.5556 42 0.7700 0.0481 0.7700 0.8775
No log 1.6296 44 0.7677 0.0481 0.7677 0.8762
No log 1.7037 46 0.7744 0.0 0.7744 0.8800
No log 1.7778 48 0.7892 0.0 0.7892 0.8884
No log 1.8519 50 0.8407 0.1050 0.8407 0.9169
No log 1.9259 52 0.9376 0.1264 0.9376 0.9683
No log 2.0 54 0.9221 0.1264 0.9221 0.9603
No log 2.0741 56 0.7818 0.2285 0.7818 0.8842
No log 2.1481 58 0.6869 0.1983 0.6869 0.8288
No log 2.2222 60 0.7223 0.0481 0.7223 0.8499
No log 2.2963 62 0.7632 0.2132 0.7632 0.8736
No log 2.3704 64 0.6797 0.1372 0.6797 0.8244
No log 2.4444 66 0.6410 0.1604 0.6410 0.8006
No log 2.5185 68 0.6390 0.2270 0.6390 0.7994
No log 2.5926 70 0.6376 0.2085 0.6376 0.7985
No log 2.6667 72 0.6776 0.3564 0.6776 0.8232
No log 2.7407 74 0.6286 0.2379 0.6286 0.7928
No log 2.8148 76 0.6215 0.2675 0.6215 0.7883
No log 2.8889 78 0.6335 0.2418 0.6335 0.7959
No log 2.9630 80 0.6532 0.2237 0.6532 0.8082
No log 3.0370 82 0.7027 0.3060 0.7027 0.8383
No log 3.1111 84 0.6886 0.2440 0.6886 0.8298
No log 3.1852 86 0.7167 0.2960 0.7167 0.8466
No log 3.2593 88 0.9161 0.2993 0.9161 0.9571
No log 3.3333 90 1.0608 0.2861 1.0608 1.0300
No log 3.4074 92 1.1054 0.3273 1.1054 1.0514
No log 3.4815 94 0.8779 0.2843 0.8779 0.9370
No log 3.5556 96 0.6992 0.4029 0.6992 0.8362
No log 3.6296 98 0.7449 0.3710 0.7449 0.8631
No log 3.7037 100 0.7742 0.3579 0.7742 0.8799
No log 3.7778 102 0.6422 0.3452 0.6422 0.8014
No log 3.8519 104 0.6555 0.4028 0.6555 0.8096
No log 3.9259 106 0.7082 0.3434 0.7082 0.8415
No log 4.0 108 0.6697 0.2857 0.6697 0.8183
No log 4.0741 110 0.6381 0.4019 0.6381 0.7988
No log 4.1481 112 0.6531 0.2973 0.6531 0.8081
No log 4.2222 114 0.6519 0.3502 0.6519 0.8074
No log 4.2963 116 0.6581 0.3170 0.6581 0.8112
No log 4.3704 118 0.6603 0.3385 0.6603 0.8126
No log 4.4444 120 0.6563 0.3228 0.6563 0.8101
No log 4.5185 122 0.6962 0.3701 0.6962 0.8344
No log 4.5926 124 0.7456 0.3099 0.7456 0.8635
No log 4.6667 126 0.6938 0.3387 0.6938 0.8329
No log 4.7407 128 0.6459 0.3385 0.6459 0.8037
No log 4.8148 130 0.6755 0.2822 0.6755 0.8219
No log 4.8889 132 0.6635 0.2822 0.6635 0.8146
No log 4.9630 134 0.6467 0.3523 0.6467 0.8041
No log 5.0370 136 0.7359 0.2632 0.7359 0.8579
No log 5.1111 138 0.7449 0.2812 0.7449 0.8631
No log 5.1852 140 0.6528 0.3649 0.6528 0.8079
No log 5.2593 142 0.6268 0.4517 0.6268 0.7917
No log 5.3333 144 0.6737 0.4776 0.6737 0.8208
No log 5.4074 146 0.6421 0.4841 0.6421 0.8013
No log 5.4815 148 0.6105 0.4838 0.6105 0.7813
No log 5.5556 150 0.6009 0.4891 0.6009 0.7752
No log 5.6296 152 0.6003 0.3445 0.6003 0.7748
No log 5.7037 154 0.6318 0.3622 0.6318 0.7949
No log 5.7778 156 0.6049 0.4086 0.6049 0.7777
No log 5.8519 158 0.6325 0.3093 0.6325 0.7953
No log 5.9259 160 0.7240 0.3140 0.7240 0.8509
No log 6.0 162 0.6810 0.4129 0.6810 0.8252
No log 6.0741 164 0.5759 0.5010 0.5759 0.7589
No log 6.1481 166 0.6097 0.4260 0.6097 0.7809
No log 6.2222 168 0.6171 0.3816 0.6171 0.7856
No log 6.2963 170 0.5880 0.4402 0.5880 0.7668
No log 6.3704 172 0.6146 0.5083 0.6146 0.7840
No log 6.4444 174 0.6840 0.4904 0.6840 0.8270
No log 6.5185 176 0.6365 0.4610 0.6365 0.7978
No log 6.5926 178 0.5855 0.3728 0.5855 0.7652
No log 6.6667 180 0.6369 0.4234 0.6369 0.7981
No log 6.7407 182 0.7276 0.4430 0.7276 0.8530
No log 6.8148 184 0.7255 0.4606 0.7255 0.8518
No log 6.8889 186 0.6470 0.4523 0.6470 0.8044
No log 6.9630 188 0.5939 0.5107 0.5939 0.7706
No log 7.0370 190 0.6002 0.5075 0.6002 0.7747
No log 7.1111 192 0.6167 0.4615 0.6167 0.7853
No log 7.1852 194 0.6364 0.5159 0.6364 0.7977
No log 7.2593 196 0.6433 0.4934 0.6433 0.8021
No log 7.3333 198 0.6541 0.4742 0.6541 0.8087
No log 7.4074 200 0.6653 0.4345 0.6653 0.8157
No log 7.4815 202 0.6546 0.4701 0.6546 0.8091
No log 7.5556 204 0.7187 0.4212 0.7187 0.8478
No log 7.6296 206 0.7244 0.4212 0.7244 0.8511
No log 7.7037 208 0.6680 0.4828 0.6680 0.8173
No log 7.7778 210 0.6629 0.4828 0.6629 0.8142
No log 7.8519 212 0.6551 0.4006 0.6551 0.8094
No log 7.9259 214 0.6699 0.3163 0.6699 0.8185
No log 8.0 216 0.6815 0.4139 0.6815 0.8255
No log 8.0741 218 0.6654 0.4096 0.6654 0.8157
No log 8.1481 220 0.6462 0.4194 0.6462 0.8039
No log 8.2222 222 0.6739 0.4301 0.6739 0.8209
No log 8.2963 224 0.6765 0.4523 0.6765 0.8225
No log 8.3704 226 0.6347 0.4423 0.6347 0.7967
No log 8.4444 228 0.6693 0.4293 0.6693 0.8181
No log 8.5185 230 0.6395 0.4444 0.6395 0.7997
No log 8.5926 232 0.6222 0.4029 0.6222 0.7888
No log 8.6667 234 0.6050 0.5042 0.6050 0.7778
No log 8.7407 236 0.5929 0.5042 0.5929 0.7700
No log 8.8148 238 0.5895 0.5114 0.5895 0.7678
No log 8.8889 240 0.5912 0.5344 0.5912 0.7689
No log 8.9630 242 0.6190 0.4555 0.6190 0.7867
No log 9.0370 244 0.6371 0.5016 0.6371 0.7982
No log 9.1111 246 0.6004 0.4681 0.6004 0.7749
No log 9.1852 248 0.5866 0.4934 0.5866 0.7659
No log 9.2593 250 0.5867 0.4441 0.5867 0.7660
No log 9.3333 252 0.5845 0.4885 0.5845 0.7645
No log 9.4074 254 0.6168 0.4614 0.6168 0.7854
No log 9.4815 256 0.6277 0.3925 0.6277 0.7923
No log 9.5556 258 0.6033 0.4044 0.6033 0.7768
No log 9.6296 260 0.6492 0.4808 0.6492 0.8057
No log 9.7037 262 0.8043 0.3934 0.8043 0.8968
No log 9.7778 264 0.9045 0.4396 0.9045 0.9511
No log 9.8519 266 0.8245 0.4152 0.8245 0.9080
No log 9.9259 268 0.7035 0.3816 0.7035 0.8387
No log 10.0 270 0.6672 0.3651 0.6672 0.8168
No log 10.0741 272 0.6834 0.3612 0.6834 0.8267
No log 10.1481 274 0.6679 0.3885 0.6679 0.8173
No log 10.2222 276 0.6497 0.3651 0.6497 0.8061
No log 10.2963 278 0.7139 0.4315 0.7139 0.8449
No log 10.3704 280 0.7092 0.4124 0.7092 0.8421
No log 10.4444 282 0.6315 0.4724 0.6315 0.7946
No log 10.5185 284 0.5877 0.5114 0.5877 0.7666
No log 10.5926 286 0.6801 0.4302 0.6801 0.8247
No log 10.6667 288 0.8359 0.3160 0.8359 0.9143
No log 10.7407 290 0.8926 0.3620 0.8926 0.9448
No log 10.8148 292 0.8187 0.3251 0.8187 0.9048
No log 10.8889 294 0.6715 0.4258 0.6715 0.8194
No log 10.9630 296 0.5878 0.5246 0.5878 0.7667
No log 11.0370 298 0.6217 0.3478 0.6217 0.7885
No log 11.1111 300 0.6787 0.3985 0.6787 0.8238
No log 11.1852 302 0.6696 0.3433 0.6696 0.8183
No log 11.2593 304 0.6390 0.3569 0.6390 0.7993
No log 11.3333 306 0.5987 0.4504 0.5987 0.7737
No log 11.4074 308 0.5926 0.4504 0.5926 0.7698
No log 11.4815 310 0.5759 0.4547 0.5759 0.7589
No log 11.5556 312 0.5728 0.4809 0.5728 0.7568
No log 11.6296 314 0.5763 0.4322 0.5763 0.7591
No log 11.7037 316 0.5728 0.4235 0.5728 0.7568
No log 11.7778 318 0.5785 0.4322 0.5785 0.7606
No log 11.8519 320 0.5814 0.4322 0.5814 0.7625
No log 11.9259 322 0.5844 0.4634 0.5844 0.7644
No log 12.0 324 0.6050 0.5326 0.6050 0.7778
No log 12.0741 326 0.6103 0.4547 0.6103 0.7812
No log 12.1481 328 0.6373 0.3806 0.6373 0.7983
No log 12.2222 330 0.6752 0.3857 0.6752 0.8217
No log 12.2963 332 0.7099 0.3470 0.7099 0.8425
No log 12.3704 334 0.6873 0.3563 0.6873 0.8290
No log 12.4444 336 0.6465 0.3780 0.6465 0.8040
No log 12.5185 338 0.6483 0.4535 0.6483 0.8052
No log 12.5926 340 0.6393 0.4934 0.6393 0.7995
No log 12.6667 342 0.6243 0.4634 0.6243 0.7901
No log 12.7407 344 0.5932 0.4299 0.5932 0.7702
No log 12.8148 346 0.5711 0.4898 0.5711 0.7557
No log 12.8889 348 0.5498 0.5361 0.5498 0.7415
No log 12.9630 350 0.5604 0.4229 0.5604 0.7486
No log 13.0370 352 0.6250 0.3746 0.6250 0.7906
No log 13.1111 354 0.6106 0.4044 0.6106 0.7814
No log 13.1852 356 0.5587 0.4504 0.5587 0.7474
No log 13.2593 358 0.5359 0.5065 0.5359 0.7320
No log 13.3333 360 0.5668 0.4198 0.5668 0.7528
No log 13.4074 362 0.5743 0.4139 0.5743 0.7578
No log 13.4815 364 0.5525 0.4838 0.5525 0.7433
No log 13.5556 366 0.5444 0.4878 0.5444 0.7378
No log 13.6296 368 0.5483 0.5022 0.5483 0.7405
No log 13.7037 370 0.5492 0.5022 0.5492 0.7411
No log 13.7778 372 0.6159 0.4354 0.6159 0.7848
No log 13.8519 374 0.7831 0.4226 0.7831 0.8850
No log 13.9259 376 0.8129 0.4305 0.8129 0.9016
No log 14.0 378 0.7415 0.4597 0.7415 0.8611
No log 14.0741 380 0.6580 0.4829 0.6580 0.8111
No log 14.1481 382 0.5741 0.4562 0.5741 0.7577
No log 14.2222 384 0.5087 0.5022 0.5087 0.7132
No log 14.2963 386 0.5031 0.4746 0.5031 0.7093
No log 14.3704 388 0.5204 0.4235 0.5204 0.7214
No log 14.4444 390 0.5262 0.4990 0.5262 0.7254
No log 14.5185 392 0.5350 0.4990 0.5350 0.7314
No log 14.5926 394 0.5384 0.4505 0.5384 0.7338
No log 14.6667 396 0.5356 0.4575 0.5356 0.7319
No log 14.7407 398 0.5349 0.4575 0.5349 0.7314
No log 14.8148 400 0.5355 0.5143 0.5355 0.7318
No log 14.8889 402 0.5587 0.5283 0.5587 0.7474
No log 14.9630 404 0.5628 0.5300 0.5628 0.7502
No log 15.0370 406 0.5512 0.4821 0.5512 0.7425
No log 15.1111 408 0.5371 0.5213 0.5371 0.7329
No log 15.1852 410 0.5524 0.5472 0.5524 0.7433
No log 15.2593 412 0.5732 0.5141 0.5732 0.7571
No log 15.3333 414 0.5894 0.4743 0.5894 0.7677
No log 15.4074 416 0.5945 0.4743 0.5945 0.7711
No log 15.4815 418 0.5906 0.4966 0.5906 0.7685
No log 15.5556 420 0.5748 0.5362 0.5748 0.7582
No log 15.6296 422 0.5639 0.5232 0.5639 0.7509
No log 15.7037 424 0.5683 0.4914 0.5683 0.7538
No log 15.7778 426 0.5661 0.5003 0.5661 0.7524
No log 15.8519 428 0.5645 0.4463 0.5645 0.7514
No log 15.9259 430 0.5674 0.4526 0.5674 0.7533
No log 16.0 432 0.5622 0.4361 0.5622 0.7498
No log 16.0741 434 0.5556 0.4361 0.5556 0.7454
No log 16.1481 436 0.5503 0.4613 0.5503 0.7418
No log 16.2222 438 0.5814 0.3865 0.5814 0.7625
No log 16.2963 440 0.6874 0.3940 0.6874 0.8291
No log 16.3704 442 0.7436 0.3461 0.7436 0.8623
No log 16.4444 444 0.6957 0.3688 0.6957 0.8341
No log 16.5185 446 0.6016 0.4888 0.6016 0.7756
No log 16.5926 448 0.5726 0.4613 0.5726 0.7567
No log 16.6667 450 0.5691 0.4613 0.5691 0.7544
No log 16.7407 452 0.5740 0.5189 0.5740 0.7576
No log 16.8148 454 0.5851 0.5003 0.5851 0.7649
No log 16.8889 456 0.5878 0.5003 0.5878 0.7667
No log 16.9630 458 0.5649 0.4361 0.5649 0.7516
No log 17.0370 460 0.5581 0.4613 0.5581 0.7471
No log 17.1111 462 0.5671 0.4934 0.5671 0.7530
No log 17.1852 464 0.5573 0.4463 0.5573 0.7465
No log 17.2593 466 0.5500 0.4361 0.5500 0.7416
No log 17.3333 468 0.5496 0.4361 0.5496 0.7413
No log 17.4074 470 0.5504 0.4361 0.5504 0.7419
No log 17.4815 472 0.5510 0.4613 0.5510 0.7423
No log 17.5556 474 0.5642 0.4768 0.5642 0.7511
No log 17.6296 476 0.5703 0.5501 0.5703 0.7552
No log 17.7037 478 0.5644 0.5232 0.5644 0.7513
No log 17.7778 480 0.5542 0.4768 0.5542 0.7444
No log 17.8519 482 0.5402 0.4526 0.5402 0.7350
No log 17.9259 484 0.5355 0.4526 0.5355 0.7318
No log 18.0 486 0.5332 0.5344 0.5332 0.7302
No log 18.0741 488 0.5247 0.5505 0.5247 0.7244
No log 18.1481 490 0.5289 0.5846 0.5289 0.7273
No log 18.2222 492 0.5750 0.4694 0.5750 0.7583
No log 18.2963 494 0.5921 0.4807 0.5921 0.7695
No log 18.3704 496 0.5355 0.5666 0.5355 0.7318
No log 18.4444 498 0.5062 0.5625 0.5062 0.7115
0.3304 18.5185 500 0.5674 0.5086 0.5674 0.7533
0.3304 18.5926 502 0.6411 0.4684 0.6411 0.8007
0.3304 18.6667 504 0.6339 0.4911 0.6339 0.7962
0.3304 18.7407 506 0.5823 0.5086 0.5823 0.7631
0.3304 18.8148 508 0.5281 0.4617 0.5281 0.7267
0.3304 18.8889 510 0.5248 0.5227 0.5248 0.7244
0.3304 18.9630 512 0.5202 0.5133 0.5202 0.7212
0.3304 19.0370 514 0.5169 0.4380 0.5169 0.7190
0.3304 19.1111 516 0.5218 0.4985 0.5218 0.7224
0.3304 19.1852 518 0.5532 0.5254 0.5532 0.7438
0.3304 19.2593 520 0.5551 0.5485 0.5551 0.7451
0.3304 19.3333 522 0.5493 0.4535 0.5493 0.7411
0.3304 19.4074 524 0.5413 0.4535 0.5413 0.7357
0.3304 19.4815 526 0.5347 0.4535 0.5348 0.7313
0.3304 19.5556 528 0.5371 0.4300 0.5371 0.7328
0.3304 19.6296 530 0.5372 0.5250 0.5372 0.7329
0.3304 19.7037 532 0.5427 0.4847 0.5427 0.7367
0.3304 19.7778 534 0.5554 0.4847 0.5554 0.7452
0.3304 19.8519 536 0.5574 0.5071 0.5574 0.7466
0.3304 19.9259 538 0.5619 0.5071 0.5619 0.7496
0.3304 20.0 540 0.5473 0.4300 0.5473 0.7398

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k10_task7_organization

Finetuned
(4023)
this model