ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k8_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0034
  • Qwk: 0.4938
  • Mse: 1.0034
  • Rmse: 1.0017

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0426 2 4.0030 -0.0169 4.0030 2.0008
No log 0.0851 4 2.0967 0.0472 2.0967 1.4480
No log 0.1277 6 1.1060 0.0495 1.1060 1.0517
No log 0.1702 8 0.7376 0.2327 0.7376 0.8589
No log 0.2128 10 0.6272 0.2247 0.6272 0.7919
No log 0.2553 12 0.6402 0.3030 0.6402 0.8001
No log 0.2979 14 0.6417 0.3287 0.6417 0.8010
No log 0.3404 16 1.1807 0.2007 1.1807 1.0866
No log 0.3830 18 1.4724 0.1910 1.4724 1.2134
No log 0.4255 20 1.0901 0.2293 1.0901 1.0441
No log 0.4681 22 0.8173 0.4797 0.8173 0.9040
No log 0.5106 24 0.8007 0.4941 0.8007 0.8948
No log 0.5532 26 0.9899 0.2848 0.9899 0.9949
No log 0.5957 28 0.9251 0.3571 0.9251 0.9618
No log 0.6383 30 0.6883 0.5011 0.6883 0.8296
No log 0.6809 32 0.6871 0.4940 0.6871 0.8289
No log 0.7234 34 0.6780 0.5037 0.6780 0.8234
No log 0.7660 36 0.6848 0.5049 0.6848 0.8276
No log 0.8085 38 0.7114 0.4866 0.7114 0.8434
No log 0.8511 40 1.0681 0.3743 1.0681 1.0335
No log 0.8936 42 1.2733 0.3402 1.2733 1.1284
No log 0.9362 44 1.1017 0.3977 1.1017 1.0496
No log 0.9787 46 0.9769 0.4639 0.9769 0.9884
No log 1.0213 48 0.7677 0.5690 0.7677 0.8762
No log 1.0638 50 0.7256 0.5686 0.7256 0.8518
No log 1.1064 52 0.9369 0.4903 0.9369 0.9679
No log 1.1489 54 1.3325 0.4057 1.3325 1.1543
No log 1.1915 56 1.1507 0.4117 1.1507 1.0727
No log 1.2340 58 0.7695 0.5341 0.7695 0.8772
No log 1.2766 60 0.6690 0.5588 0.6690 0.8179
No log 1.3191 62 0.8540 0.4654 0.8540 0.9241
No log 1.3617 64 0.8300 0.4637 0.8300 0.9110
No log 1.4043 66 0.7587 0.5755 0.7587 0.8711
No log 1.4468 68 0.8229 0.5648 0.8229 0.9071
No log 1.4894 70 0.9002 0.5633 0.9002 0.9488
No log 1.5319 72 0.9417 0.5579 0.9417 0.9704
No log 1.5745 74 0.9437 0.5568 0.9437 0.9714
No log 1.6170 76 0.9388 0.5300 0.9388 0.9689
No log 1.6596 78 0.9052 0.5521 0.9052 0.9514
No log 1.7021 80 0.8390 0.5365 0.8390 0.9160
No log 1.7447 82 0.8136 0.5237 0.8136 0.9020
No log 1.7872 84 0.7595 0.5294 0.7595 0.8715
No log 1.8298 86 0.7749 0.5516 0.7749 0.8803
No log 1.8723 88 0.7648 0.5323 0.7648 0.8745
No log 1.9149 90 1.0041 0.4359 1.0041 1.0020
No log 1.9574 92 1.2178 0.3521 1.2178 1.1036
No log 2.0 94 1.0817 0.4076 1.0817 1.0400
No log 2.0426 96 0.7890 0.5099 0.7890 0.8883
No log 2.0851 98 0.8507 0.5669 0.8507 0.9223
No log 2.1277 100 1.1011 0.4818 1.1011 1.0494
No log 2.1702 102 0.9920 0.4846 0.9920 0.9960
No log 2.2128 104 0.7750 0.5625 0.7750 0.8803
No log 2.2553 106 0.7000 0.5946 0.7000 0.8367
No log 2.2979 108 0.7107 0.6117 0.7107 0.8430
No log 2.3404 110 0.7774 0.5627 0.7774 0.8817
No log 2.3830 112 0.8215 0.5868 0.8215 0.9064
No log 2.4255 114 0.9167 0.5442 0.9167 0.9574
No log 2.4681 116 1.0921 0.5245 1.0921 1.0450
No log 2.5106 118 1.1327 0.5115 1.1327 1.0643
No log 2.5532 120 1.1283 0.5238 1.1283 1.0622
No log 2.5957 122 1.0888 0.5267 1.0888 1.0434
No log 2.6383 124 1.0371 0.4977 1.0371 1.0184
No log 2.6809 126 1.0548 0.4952 1.0548 1.0271
No log 2.7234 128 1.0060 0.5162 1.0060 1.0030
No log 2.7660 130 0.9886 0.5348 0.9886 0.9943
No log 2.8085 132 1.0189 0.5121 1.0189 1.0094
No log 2.8511 134 1.1054 0.4615 1.1054 1.0514
No log 2.8936 136 1.0382 0.4657 1.0382 1.0189
No log 2.9362 138 1.0461 0.4559 1.0461 1.0228
No log 2.9787 140 1.0583 0.4468 1.0583 1.0287
No log 3.0213 142 1.0225 0.4430 1.0225 1.0112
No log 3.0638 144 0.9742 0.5185 0.9742 0.9870
No log 3.1064 146 0.9798 0.5074 0.9798 0.9898
No log 3.1489 148 0.9865 0.4917 0.9865 0.9932
No log 3.1915 150 1.0352 0.4417 1.0352 1.0174
No log 3.2340 152 0.9943 0.4643 0.9943 0.9972
No log 3.2766 154 0.9298 0.4878 0.9298 0.9643
No log 3.3191 156 0.8871 0.4900 0.8871 0.9419
No log 3.3617 158 0.8354 0.5331 0.8354 0.9140
No log 3.4043 160 0.8822 0.5310 0.8822 0.9393
No log 3.4468 162 0.9611 0.5329 0.9611 0.9803
No log 3.4894 164 1.0452 0.5459 1.0452 1.0223
No log 3.5319 166 1.1121 0.4478 1.1121 1.0546
No log 3.5745 168 1.2082 0.4536 1.2082 1.0992
No log 3.6170 170 1.2337 0.4577 1.2337 1.1107
No log 3.6596 172 1.2114 0.4565 1.2114 1.1006
No log 3.7021 174 1.2312 0.5031 1.2312 1.1096
No log 3.7447 176 1.2118 0.5070 1.2118 1.1008
No log 3.7872 178 1.0741 0.5434 1.0741 1.0364
No log 3.8298 180 1.0029 0.5162 1.0029 1.0015
No log 3.8723 182 0.9576 0.4724 0.9576 0.9786
No log 3.9149 184 0.9601 0.4763 0.9601 0.9799
No log 3.9574 186 1.0051 0.4734 1.0051 1.0025
No log 4.0 188 1.0376 0.4577 1.0376 1.0186
No log 4.0426 190 1.0685 0.4698 1.0685 1.0337
No log 4.0851 192 1.1065 0.5008 1.1065 1.0519
No log 4.1277 194 1.0499 0.4864 1.0499 1.0247
No log 4.1702 196 0.9754 0.4978 0.9754 0.9876
No log 4.2128 198 0.9294 0.4708 0.9294 0.9641
No log 4.2553 200 0.8615 0.5190 0.8615 0.9282
No log 4.2979 202 0.8562 0.5339 0.8562 0.9253
No log 4.3404 204 0.8997 0.5729 0.8997 0.9485
No log 4.3830 206 0.9387 0.5256 0.9387 0.9689
No log 4.4255 208 0.9445 0.5221 0.9445 0.9718
No log 4.4681 210 0.9986 0.5244 0.9986 0.9993
No log 4.5106 212 1.1187 0.4753 1.1187 1.0577
No log 4.5532 214 1.1991 0.4761 1.1991 1.0950
No log 4.5957 216 1.2760 0.4564 1.2760 1.1296
No log 4.6383 218 1.3071 0.4729 1.3071 1.1433
No log 4.6809 220 1.3614 0.4385 1.3614 1.1668
No log 4.7234 222 1.3491 0.4598 1.3491 1.1615
No log 4.7660 224 1.2696 0.4715 1.2696 1.1268
No log 4.8085 226 1.2517 0.4538 1.2517 1.1188
No log 4.8511 228 1.1857 0.4460 1.1857 1.0889
No log 4.8936 230 1.0434 0.4522 1.0434 1.0215
No log 4.9362 232 0.9010 0.5278 0.9010 0.9492
No log 4.9787 234 0.8775 0.5414 0.8775 0.9367
No log 5.0213 236 0.9220 0.5320 0.9220 0.9602
No log 5.0638 238 0.9578 0.5402 0.9578 0.9787
No log 5.1064 240 0.9718 0.5388 0.9718 0.9858
No log 5.1489 242 0.9375 0.5365 0.9375 0.9683
No log 5.1915 244 0.8957 0.5315 0.8957 0.9464
No log 5.2340 246 0.8688 0.5353 0.8688 0.9321
No log 5.2766 248 0.8968 0.5271 0.8968 0.9470
No log 5.3191 250 0.9053 0.5453 0.9053 0.9515
No log 5.3617 252 0.9405 0.5440 0.9405 0.9698
No log 5.4043 254 0.9103 0.5394 0.9103 0.9541
No log 5.4468 256 0.8751 0.5185 0.8751 0.9354
No log 5.4894 258 0.8935 0.5528 0.8935 0.9453
No log 5.5319 260 0.9130 0.4850 0.9130 0.9555
No log 5.5745 262 0.8961 0.4988 0.8961 0.9466
No log 5.6170 264 0.8735 0.5129 0.8735 0.9346
No log 5.6596 266 0.8635 0.5351 0.8635 0.9292
No log 5.7021 268 0.8551 0.5099 0.8551 0.9247
No log 5.7447 270 0.8808 0.4743 0.8808 0.9385
No log 5.7872 272 0.8860 0.4660 0.8860 0.9412
No log 5.8298 274 0.8494 0.4980 0.8494 0.9216
No log 5.8723 276 0.8398 0.5216 0.8398 0.9164
No log 5.9149 278 0.8371 0.5469 0.8371 0.9149
No log 5.9574 280 0.8562 0.5168 0.8562 0.9253
No log 6.0 282 0.8789 0.5423 0.8789 0.9375
No log 6.0426 284 0.9296 0.5278 0.9296 0.9642
No log 6.0851 286 1.0229 0.4895 1.0229 1.0114
No log 6.1277 288 1.0674 0.4909 1.0674 1.0331
No log 6.1702 290 1.0150 0.4971 1.0150 1.0075
No log 6.2128 292 0.9439 0.5150 0.9439 0.9715
No log 6.2553 294 0.9167 0.5342 0.9167 0.9575
No log 6.2979 296 0.8908 0.5333 0.8908 0.9438
No log 6.3404 298 0.8969 0.5080 0.8969 0.9471
No log 6.3830 300 0.9027 0.5125 0.9027 0.9501
No log 6.4255 302 0.8833 0.5211 0.8833 0.9399
No log 6.4681 304 0.8950 0.4975 0.8950 0.9461
No log 6.5106 306 0.8949 0.4975 0.8949 0.9460
No log 6.5532 308 0.8939 0.5223 0.8939 0.9455
No log 6.5957 310 0.9028 0.5111 0.9028 0.9501
No log 6.6383 312 0.8999 0.5120 0.8999 0.9486
No log 6.6809 314 0.9073 0.5166 0.9073 0.9525
No log 6.7234 316 0.9224 0.4945 0.9224 0.9604
No log 6.7660 318 0.9604 0.4916 0.9604 0.9800
No log 6.8085 320 1.0126 0.4629 1.0126 1.0063
No log 6.8511 322 1.0037 0.4608 1.0037 1.0019
No log 6.8936 324 0.9652 0.4820 0.9652 0.9824
No log 6.9362 326 0.9472 0.5347 0.9472 0.9732
No log 6.9787 328 0.9573 0.5091 0.9573 0.9784
No log 7.0213 330 0.9433 0.5206 0.9433 0.9712
No log 7.0638 332 0.8912 0.5282 0.8912 0.9440
No log 7.1064 334 0.8453 0.5208 0.8453 0.9194
No log 7.1489 336 0.8316 0.5206 0.8316 0.9119
No log 7.1915 338 0.8437 0.5472 0.8437 0.9185
No log 7.2340 340 0.8692 0.5332 0.8692 0.9323
No log 7.2766 342 0.8983 0.5261 0.8983 0.9478
No log 7.3191 344 0.9391 0.5032 0.9391 0.9691
No log 7.3617 346 0.9628 0.4693 0.9628 0.9812
No log 7.4043 348 0.9699 0.4704 0.9699 0.9848
No log 7.4468 350 0.9772 0.4955 0.9772 0.9885
No log 7.4894 352 0.9852 0.4918 0.9852 0.9926
No log 7.5319 354 0.9923 0.4918 0.9923 0.9961
No log 7.5745 356 1.0042 0.4736 1.0042 1.0021
No log 7.6170 358 1.0174 0.4545 1.0174 1.0087
No log 7.6596 360 1.0468 0.4625 1.0468 1.0232
No log 7.7021 362 1.0571 0.4695 1.0571 1.0281
No log 7.7447 364 1.0627 0.4863 1.0627 1.0309
No log 7.7872 366 1.0800 0.5257 1.0800 1.0392
No log 7.8298 368 1.0718 0.5340 1.0718 1.0353
No log 7.8723 370 1.0502 0.5204 1.0502 1.0248
No log 7.9149 372 1.0194 0.5074 1.0194 1.0097
No log 7.9574 374 1.0060 0.5165 1.0060 1.0030
No log 8.0 376 0.9883 0.5165 0.9883 0.9941
No log 8.0426 378 0.9649 0.5132 0.9649 0.9823
No log 8.0851 380 0.9436 0.5038 0.9436 0.9714
No log 8.1277 382 0.9326 0.5078 0.9326 0.9657
No log 8.1702 384 0.9320 0.5033 0.9320 0.9654
No log 8.2128 386 0.9355 0.4885 0.9355 0.9672
No log 8.2553 388 0.9450 0.4918 0.9450 0.9721
No log 8.2979 390 0.9647 0.4912 0.9647 0.9822
No log 8.3404 392 0.9861 0.5072 0.9861 0.9930
No log 8.3830 394 1.0181 0.4943 1.0181 1.0090
No log 8.4255 396 1.0474 0.4975 1.0474 1.0234
No log 8.4681 398 1.0518 0.4975 1.0518 1.0256
No log 8.5106 400 1.0519 0.4975 1.0519 1.0256
No log 8.5532 402 1.0441 0.4695 1.0441 1.0218
No log 8.5957 404 1.0264 0.4834 1.0264 1.0131
No log 8.6383 406 1.0156 0.4909 1.0156 1.0078
No log 8.6809 408 1.0153 0.4725 1.0153 1.0076
No log 8.7234 410 1.0088 0.4777 1.0088 1.0044
No log 8.7660 412 0.9979 0.494 0.9979 0.9989
No log 8.8085 414 0.9955 0.4855 0.9955 0.9978
No log 8.8511 416 0.9919 0.4773 0.9919 0.9959
No log 8.8936 418 0.9882 0.475 0.9882 0.9941
No log 8.9362 420 1.0003 0.5037 1.0003 1.0001
No log 8.9787 422 1.0075 0.5117 1.0075 1.0038
No log 9.0213 424 1.0133 0.5079 1.0133 1.0066
No log 9.0638 426 1.0213 0.4890 1.0213 1.0106
No log 9.1064 428 1.0205 0.4797 1.0205 1.0102
No log 9.1489 430 1.0182 0.4797 1.0182 1.0091
No log 9.1915 432 1.0154 0.4754 1.0154 1.0077
No log 9.2340 434 1.0098 0.4863 1.0098 1.0049
No log 9.2766 436 1.0012 0.4969 1.0012 1.0006
No log 9.3191 438 0.9944 0.4930 0.9944 0.9972
No log 9.3617 440 0.9927 0.4930 0.9927 0.9963
No log 9.4043 442 0.9897 0.4949 0.9897 0.9948
No log 9.4468 444 0.9861 0.4949 0.9861 0.9930
No log 9.4894 446 0.9865 0.4949 0.9865 0.9932
No log 9.5319 448 0.9878 0.4860 0.9878 0.9939
No log 9.5745 450 0.9871 0.4868 0.9871 0.9935
No log 9.6170 452 0.9892 0.4890 0.9892 0.9946
No log 9.6596 454 0.9920 0.4890 0.9920 0.9960
No log 9.7021 456 0.9933 0.4890 0.9933 0.9966
No log 9.7447 458 0.9946 0.4890 0.9946 0.9973
No log 9.7872 460 0.9956 0.4890 0.9956 0.9978
No log 9.8298 462 0.9977 0.4868 0.9977 0.9988
No log 9.8723 464 0.9991 0.4860 0.9991 0.9996
No log 9.9149 466 1.0014 0.4938 1.0014 1.0007
No log 9.9574 468 1.0028 0.4938 1.0028 1.0014
No log 10.0 470 1.0034 0.4938 1.0034 1.0017

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k8_task2_organization

Finetuned
(4023)
this model