ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k7_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9957
  • Qwk: 0.5082
  • Mse: 0.9957
  • Rmse: 0.9979

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 3.8414 -0.0004 3.8414 1.9599
No log 0.1 4 1.7535 0.1246 1.7535 1.3242
No log 0.15 6 1.1221 0.1359 1.1221 1.0593
No log 0.2 8 0.8593 0.0750 0.8593 0.9270
No log 0.25 10 0.7497 0.0875 0.7497 0.8658
No log 0.3 12 0.6830 0.2159 0.6830 0.8264
No log 0.35 14 0.7104 0.1436 0.7104 0.8428
No log 0.4 16 0.7174 0.1615 0.7174 0.8470
No log 0.45 18 0.6506 0.2713 0.6506 0.8066
No log 0.5 20 0.6061 0.2333 0.6061 0.7785
No log 0.55 22 0.5620 0.2645 0.5620 0.7497
No log 0.6 24 0.5330 0.4146 0.5330 0.7301
No log 0.65 26 0.5203 0.4163 0.5203 0.7213
No log 0.7 28 0.5079 0.4876 0.5079 0.7127
No log 0.75 30 0.5140 0.4931 0.5140 0.7169
No log 0.8 32 0.5221 0.5279 0.5221 0.7226
No log 0.85 34 0.6252 0.5153 0.6252 0.7907
No log 0.9 36 0.9745 0.3891 0.9745 0.9872
No log 0.95 38 0.7182 0.4924 0.7182 0.8475
No log 1.0 40 0.5644 0.5574 0.5644 0.7512
No log 1.05 42 0.5643 0.5631 0.5643 0.7512
No log 1.1 44 0.6120 0.5529 0.6120 0.7823
No log 1.15 46 0.5998 0.5593 0.5998 0.7744
No log 1.2 48 0.5729 0.5467 0.5729 0.7569
No log 1.25 50 0.5772 0.5308 0.5772 0.7597
No log 1.3 52 0.6095 0.5202 0.6095 0.7807
No log 1.35 54 0.6186 0.5552 0.6186 0.7865
No log 1.4 56 0.6580 0.5610 0.6580 0.8112
No log 1.45 58 0.7026 0.5579 0.7026 0.8382
No log 1.5 60 0.7892 0.5548 0.7892 0.8884
No log 1.55 62 0.7894 0.5847 0.7894 0.8885
No log 1.6 64 0.7658 0.5749 0.7658 0.8751
No log 1.65 66 0.7906 0.5696 0.7906 0.8891
No log 1.7 68 0.8359 0.5741 0.8359 0.9143
No log 1.75 70 0.9797 0.5587 0.9797 0.9898
No log 1.8 72 1.0481 0.5358 1.0481 1.0238
No log 1.85 74 0.7749 0.5409 0.7749 0.8803
No log 1.9 76 0.6797 0.5349 0.6797 0.8244
No log 1.95 78 0.6943 0.5349 0.6943 0.8332
No log 2.0 80 0.6724 0.5702 0.6724 0.8200
No log 2.05 82 0.7605 0.5828 0.7605 0.8721
No log 2.1 84 1.3588 0.4634 1.3588 1.1657
No log 2.15 86 1.5990 0.4052 1.5990 1.2645
No log 2.2 88 1.2584 0.4813 1.2584 1.1218
No log 2.25 90 0.7593 0.5958 0.7593 0.8714
No log 2.3 92 0.7420 0.5508 0.7420 0.8614
No log 2.35 94 0.7154 0.5652 0.7154 0.8458
No log 2.4 96 0.6466 0.5887 0.6466 0.8041
No log 2.45 98 0.8131 0.5709 0.8131 0.9017
No log 2.5 100 0.9397 0.5542 0.9397 0.9694
No log 2.55 102 0.9921 0.5455 0.9921 0.9961
No log 2.6 104 0.8991 0.5348 0.8991 0.9482
No log 2.65 106 0.9380 0.5378 0.9380 0.9685
No log 2.7 108 0.9466 0.5487 0.9466 0.9730
No log 2.75 110 0.7382 0.5483 0.7382 0.8592
No log 2.8 112 0.8953 0.4497 0.8953 0.9462
No log 2.85 114 1.2332 0.3295 1.2332 1.1105
No log 2.9 116 1.1457 0.3762 1.1457 1.0704
No log 2.95 118 0.8903 0.4936 0.8903 0.9436
No log 3.0 120 1.1135 0.4933 1.1135 1.0552
No log 3.05 122 1.3199 0.4559 1.3199 1.1489
No log 3.1 124 1.2982 0.4490 1.2982 1.1394
No log 3.15 126 1.0380 0.4729 1.0380 1.0188
No log 3.2 128 0.8737 0.5447 0.8737 0.9347
No log 3.25 130 0.9272 0.4753 0.9272 0.9629
No log 3.3 132 0.9208 0.4508 0.9208 0.9596
No log 3.35 134 0.7907 0.4950 0.7907 0.8892
No log 3.4 136 0.7992 0.5575 0.7992 0.8940
No log 3.45 138 1.0994 0.4853 1.0994 1.0485
No log 3.5 140 1.1707 0.4649 1.1707 1.0820
No log 3.55 142 0.9968 0.4886 0.9968 0.9984
No log 3.6 144 0.8862 0.5492 0.8862 0.9414
No log 3.65 146 0.8942 0.5459 0.8942 0.9456
No log 3.7 148 1.0412 0.4743 1.0412 1.0204
No log 3.75 150 1.2145 0.4622 1.2145 1.1020
No log 3.8 152 1.1521 0.4718 1.1521 1.0734
No log 3.85 154 1.0651 0.4855 1.0651 1.0321
No log 3.9 156 0.9901 0.4714 0.9901 0.9950
No log 3.95 158 0.9822 0.4745 0.9822 0.9910
No log 4.0 160 0.9938 0.4752 0.9938 0.9969
No log 4.05 162 1.0830 0.4605 1.0830 1.0407
No log 4.1 164 1.3323 0.4628 1.3323 1.1543
No log 4.15 166 1.4798 0.4418 1.4798 1.2165
No log 4.2 168 1.3385 0.4724 1.3385 1.1569
No log 4.25 170 1.0257 0.5061 1.0257 1.0127
No log 4.3 172 0.8813 0.5554 0.8813 0.9388
No log 4.35 174 0.8310 0.5592 0.8310 0.9116
No log 4.4 176 0.8450 0.5264 0.8450 0.9192
No log 4.45 178 0.8431 0.5027 0.8431 0.9182
No log 4.5 180 0.9215 0.4756 0.9215 0.9599
No log 4.55 182 1.1347 0.4617 1.1347 1.0652
No log 4.6 184 1.2375 0.4584 1.2375 1.1124
No log 4.65 186 1.1086 0.4414 1.1086 1.0529
No log 4.7 188 0.9470 0.4756 0.9470 0.9731
No log 4.75 190 0.8405 0.5453 0.8405 0.9168
No log 4.8 192 0.8368 0.5761 0.8368 0.9148
No log 4.85 194 0.8819 0.5598 0.8819 0.9391
No log 4.9 196 1.0677 0.4782 1.0677 1.0333
No log 4.95 198 1.2343 0.4578 1.2343 1.1110
No log 5.0 200 1.2901 0.4658 1.2901 1.1358
No log 5.05 202 1.1682 0.4482 1.1682 1.0808
No log 5.1 204 0.9759 0.5082 0.9759 0.9879
No log 5.15 206 0.9232 0.5245 0.9232 0.9608
No log 5.2 208 0.9029 0.5174 0.9029 0.9502
No log 5.25 210 0.9162 0.5142 0.9162 0.9572
No log 5.3 212 0.9207 0.4814 0.9207 0.9595
No log 5.35 214 0.9776 0.4987 0.9776 0.9888
No log 5.4 216 1.0946 0.4675 1.0946 1.0462
No log 5.45 218 1.0937 0.4675 1.0937 1.0458
No log 5.5 220 0.9645 0.4556 0.9645 0.9821
No log 5.55 222 0.8801 0.5274 0.8801 0.9382
No log 5.6 224 0.8741 0.5175 0.8741 0.9349
No log 5.65 226 0.8925 0.5158 0.8925 0.9447
No log 5.7 228 0.9125 0.5294 0.9125 0.9553
No log 5.75 230 0.9377 0.5139 0.9377 0.9684
No log 5.8 232 1.0149 0.4992 1.0149 1.0074
No log 5.85 234 1.1244 0.4855 1.1244 1.0604
No log 5.9 236 1.2971 0.4479 1.2971 1.1389
No log 5.95 238 1.3821 0.4545 1.3821 1.1756
No log 6.0 240 1.3742 0.4420 1.3742 1.1723
No log 6.05 242 1.3104 0.4553 1.3104 1.1447
No log 6.1 244 1.2469 0.4672 1.2469 1.1166
No log 6.15 246 1.2066 0.4647 1.2066 1.0984
No log 6.2 248 1.1203 0.4624 1.1203 1.0584
No log 6.25 250 1.0773 0.4732 1.0773 1.0379
No log 6.3 252 1.0497 0.5176 1.0497 1.0245
No log 6.35 254 1.0367 0.5140 1.0367 1.0182
No log 6.4 256 1.0390 0.4739 1.0390 1.0193
No log 6.45 258 1.0249 0.4622 1.0249 1.0124
No log 6.5 260 1.1264 0.4863 1.1264 1.0613
No log 6.55 262 1.2807 0.4480 1.2807 1.1317
No log 6.6 264 1.3708 0.4435 1.3708 1.1708
No log 6.65 266 1.3242 0.4540 1.3242 1.1507
No log 6.7 268 1.1589 0.4843 1.1589 1.0765
No log 6.75 270 1.0310 0.4784 1.0310 1.0154
No log 6.8 272 0.9804 0.4880 0.9804 0.9902
No log 6.85 274 0.9788 0.5104 0.9788 0.9893
No log 6.9 276 0.9904 0.4899 0.9904 0.9952
No log 6.95 278 1.0043 0.4878 1.0043 1.0021
No log 7.0 280 1.0405 0.4926 1.0405 1.0201
No log 7.05 282 1.0639 0.4896 1.0639 1.0315
No log 7.1 284 1.1201 0.4736 1.1201 1.0583
No log 7.15 286 1.1298 0.4786 1.1298 1.0629
No log 7.2 288 1.0615 0.4821 1.0615 1.0303
No log 7.25 290 0.9548 0.4935 0.9548 0.9772
No log 7.3 292 0.8949 0.5080 0.8949 0.9460
No log 7.35 294 0.8768 0.5292 0.8768 0.9364
No log 7.4 296 0.9050 0.5043 0.9050 0.9513
No log 7.45 298 1.0012 0.51 1.0012 1.0006
No log 7.5 300 1.0626 0.4975 1.0626 1.0308
No log 7.55 302 1.0457 0.4849 1.0457 1.0226
No log 7.6 304 0.9984 0.4958 0.9984 0.9992
No log 7.65 306 0.9656 0.5133 0.9656 0.9826
No log 7.7 308 0.9399 0.5091 0.9399 0.9695
No log 7.75 310 0.8961 0.5294 0.8961 0.9466
No log 7.8 312 0.8930 0.5334 0.8930 0.9450
No log 7.85 314 0.8815 0.5228 0.8815 0.9389
No log 7.9 316 0.8743 0.5204 0.8743 0.9351
No log 7.95 318 0.9048 0.5186 0.9048 0.9512
No log 8.0 320 0.9422 0.5054 0.9422 0.9707
No log 8.05 322 1.0185 0.5125 1.0185 1.0092
No log 8.1 324 1.0586 0.4957 1.0586 1.0289
No log 8.15 326 1.0837 0.4941 1.0837 1.0410
No log 8.2 328 1.0793 0.4994 1.0793 1.0389
No log 8.25 330 1.0257 0.5166 1.0257 1.0128
No log 8.3 332 0.9697 0.4747 0.9697 0.9847
No log 8.35 334 0.9388 0.4914 0.9388 0.9689
No log 8.4 336 0.9215 0.4914 0.9215 0.9599
No log 8.45 338 0.9296 0.4914 0.9296 0.9641
No log 8.5 340 0.9469 0.4867 0.9469 0.9731
No log 8.55 342 0.9954 0.4925 0.9954 0.9977
No log 8.6 344 1.0727 0.4902 1.0727 1.0357
No log 8.65 346 1.1384 0.4808 1.1384 1.0670
No log 8.7 348 1.1549 0.4645 1.1549 1.0746
No log 8.75 350 1.1313 0.4808 1.1313 1.0636
No log 8.8 352 1.0933 0.4831 1.0933 1.0456
No log 8.85 354 1.0697 0.5 1.0697 1.0343
No log 8.9 356 1.0318 0.4784 1.0318 1.0158
No log 8.95 358 1.0038 0.4849 1.0038 1.0019
No log 9.0 360 0.9919 0.4936 0.9919 0.9959
No log 9.05 362 0.9976 0.5200 0.9976 0.9988
No log 9.1 364 1.0081 0.5200 1.0081 1.0040
No log 9.15 366 1.0131 0.5200 1.0131 1.0065
No log 9.2 368 1.0194 0.4986 1.0194 1.0097
No log 9.25 370 1.0228 0.4986 1.0228 1.0114
No log 9.3 372 1.0190 0.4933 1.0190 1.0095
No log 9.35 374 1.0127 0.4985 1.0127 1.0063
No log 9.4 376 1.0093 0.4985 1.0093 1.0047
No log 9.45 378 1.0105 0.4836 1.0105 1.0053
No log 9.5 380 1.0162 0.4859 1.0162 1.0081
No log 9.55 382 1.0120 0.4893 1.0120 1.0060
No log 9.6 384 1.0079 0.4893 1.0079 1.0039
No log 9.65 386 1.0038 0.4893 1.0038 1.0019
No log 9.7 388 1.0000 0.4849 1.0000 1.0000
No log 9.75 390 0.9998 0.4891 0.9998 0.9999
No log 9.8 392 1.0002 0.4891 1.0002 1.0001
No log 9.85 394 1.0004 0.5123 1.0004 1.0002
No log 9.9 396 0.9980 0.5082 0.9980 0.9990
No log 9.95 398 0.9966 0.5082 0.9966 0.9983
No log 10.0 400 0.9957 0.5082 0.9957 0.9979

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k7_task2_organization

Finetuned
(4023)
this model