ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k6_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9791
  • Qwk: 0.3696
  • Mse: 0.9791
  • Rmse: 0.9895

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0606 2 4.1307 0.0249 4.1307 2.0324
No log 0.1212 4 2.3083 0.0845 2.3083 1.5193
No log 0.1818 6 1.2185 0.0864 1.2185 1.1039
No log 0.2424 8 0.8098 0.1001 0.8098 0.8999
No log 0.3030 10 0.6746 0.2370 0.6746 0.8213
No log 0.3636 12 0.7091 0.2014 0.7091 0.8421
No log 0.4242 14 0.7181 0.1971 0.7181 0.8474
No log 0.4848 16 0.6990 0.2108 0.6990 0.8361
No log 0.5455 18 0.6956 0.2326 0.6956 0.8340
No log 0.6061 20 0.6554 0.2318 0.6554 0.8096
No log 0.6667 22 0.6432 0.2281 0.6432 0.8020
No log 0.7273 24 0.6199 0.1984 0.6199 0.7874
No log 0.7879 26 0.6248 0.2326 0.6248 0.7904
No log 0.8485 28 0.6090 0.3314 0.6090 0.7804
No log 0.9091 30 0.5909 0.4085 0.5909 0.7687
No log 0.9697 32 0.5855 0.3365 0.5855 0.7652
No log 1.0303 34 0.6023 0.3705 0.6023 0.7761
No log 1.0909 36 0.6196 0.4107 0.6196 0.7872
No log 1.1515 38 0.5732 0.3848 0.5732 0.7571
No log 1.2121 40 0.5686 0.3429 0.5686 0.7540
No log 1.2727 42 0.5600 0.3824 0.5600 0.7483
No log 1.3333 44 0.6049 0.4549 0.6049 0.7778
No log 1.3939 46 0.7023 0.3943 0.7023 0.8381
No log 1.4545 48 0.7869 0.3709 0.7869 0.8871
No log 1.5152 50 0.9962 0.3309 0.9962 0.9981
No log 1.5758 52 1.0908 0.3486 1.0908 1.0444
No log 1.6364 54 1.0132 0.3840 1.0132 1.0066
No log 1.6970 56 0.9323 0.3914 0.9323 0.9656
No log 1.7576 58 0.8469 0.3934 0.8469 0.9203
No log 1.8182 60 0.9207 0.3654 0.9207 0.9595
No log 1.8788 62 1.0407 0.3726 1.0407 1.0201
No log 1.9394 64 1.0650 0.3678 1.0650 1.0320
No log 2.0 66 1.1383 0.3611 1.1383 1.0669
No log 2.0606 68 1.1288 0.3587 1.1288 1.0624
No log 2.1212 70 1.1213 0.3892 1.1213 1.0589
No log 2.1818 72 1.1993 0.4071 1.1993 1.0951
No log 2.2424 74 1.2944 0.3878 1.2944 1.1377
No log 2.3030 76 1.3960 0.3521 1.3960 1.1815
No log 2.3636 78 1.2647 0.3789 1.2647 1.1246
No log 2.4242 80 1.2162 0.3881 1.2162 1.1028
No log 2.4848 82 1.2669 0.3841 1.2669 1.1255
No log 2.5455 84 1.1465 0.3857 1.1465 1.0708
No log 2.6061 86 1.1507 0.3835 1.1507 1.0727
No log 2.6667 88 1.4169 0.3012 1.4169 1.1903
No log 2.7273 90 1.4889 0.3197 1.4889 1.2202
No log 2.7879 92 1.4851 0.3289 1.4851 1.2186
No log 2.8485 94 1.4873 0.2729 1.4873 1.2196
No log 2.9091 96 1.2201 0.3239 1.2201 1.1046
No log 2.9697 98 1.0043 0.3779 1.0043 1.0021
No log 3.0303 100 0.9065 0.4250 0.9065 0.9521
No log 3.0909 102 0.8769 0.4197 0.8769 0.9364
No log 3.1515 104 0.9468 0.4291 0.9468 0.9730
No log 3.2121 106 1.0598 0.3755 1.0598 1.0294
No log 3.2727 108 1.1153 0.3735 1.1153 1.0561
No log 3.3333 110 1.0235 0.3826 1.0235 1.0117
No log 3.3939 112 0.9421 0.3957 0.9421 0.9706
No log 3.4545 114 0.9440 0.4020 0.9440 0.9716
No log 3.5152 116 0.9323 0.4092 0.9323 0.9655
No log 3.5758 118 0.9474 0.4436 0.9474 0.9733
No log 3.6364 120 0.9950 0.4226 0.9950 0.9975
No log 3.6970 122 1.0719 0.3838 1.0719 1.0353
No log 3.7576 124 1.2666 0.4046 1.2666 1.1254
No log 3.8182 126 1.4116 0.4303 1.4116 1.1881
No log 3.8788 128 1.4931 0.4207 1.4931 1.2219
No log 3.9394 130 1.3472 0.4189 1.3472 1.1607
No log 4.0 132 1.2379 0.4172 1.2379 1.1126
No log 4.0606 134 1.2003 0.4059 1.2003 1.0956
No log 4.1212 136 1.0669 0.4134 1.0669 1.0329
No log 4.1818 138 0.9810 0.3694 0.9810 0.9905
No log 4.2424 140 0.8620 0.3774 0.8620 0.9284
No log 4.3030 142 0.8216 0.3776 0.8216 0.9064
No log 4.3636 144 0.8485 0.3766 0.8485 0.9212
No log 4.4242 146 0.9814 0.3973 0.9814 0.9907
No log 4.4848 148 1.0780 0.4050 1.0780 1.0383
No log 4.5455 150 1.1209 0.3737 1.1209 1.0587
No log 4.6061 152 1.1379 0.3775 1.1379 1.0667
No log 4.6667 154 1.1671 0.3882 1.1671 1.0803
No log 4.7273 156 1.2671 0.3814 1.2671 1.1257
No log 4.7879 158 1.3806 0.3943 1.3806 1.1750
No log 4.8485 160 1.2917 0.3925 1.2917 1.1365
No log 4.9091 162 1.0953 0.3774 1.0953 1.0466
No log 4.9697 164 0.8896 0.4256 0.8896 0.9432
No log 5.0303 166 0.8209 0.4096 0.8209 0.9060
No log 5.0909 168 0.8281 0.4040 0.8281 0.9100
No log 5.1515 170 0.9049 0.4067 0.9049 0.9513
No log 5.2121 172 1.0835 0.4107 1.0835 1.0409
No log 5.2727 174 1.2358 0.4101 1.2358 1.1117
No log 5.3333 176 1.1966 0.3942 1.1966 1.0939
No log 5.3939 178 1.1597 0.4136 1.1597 1.0769
No log 5.4545 180 1.1234 0.4312 1.1234 1.0599
No log 5.5152 182 1.0893 0.3983 1.0893 1.0437
No log 5.5758 184 1.0634 0.4017 1.0634 1.0312
No log 5.6364 186 1.0146 0.4052 1.0146 1.0073
No log 5.6970 188 0.9412 0.4326 0.9412 0.9701
No log 5.7576 190 0.9333 0.4207 0.9333 0.9661
No log 5.8182 192 0.9558 0.4066 0.9558 0.9776
No log 5.8788 194 0.9201 0.3794 0.9201 0.9592
No log 5.9394 196 0.8730 0.4071 0.8730 0.9343
No log 6.0 198 0.8285 0.4213 0.8285 0.9102
No log 6.0606 200 0.8207 0.4590 0.8207 0.9059
No log 6.1212 202 0.8589 0.4386 0.8589 0.9268
No log 6.1818 204 0.9429 0.4401 0.9429 0.9710
No log 6.2424 206 1.0881 0.3980 1.0881 1.0431
No log 6.3030 208 1.1993 0.4023 1.1993 1.0951
No log 6.3636 210 1.2751 0.3934 1.2751 1.1292
No log 6.4242 212 1.2756 0.3631 1.2756 1.1294
No log 6.4848 214 1.2118 0.3973 1.2118 1.1008
No log 6.5455 216 1.1944 0.3984 1.1944 1.0929
No log 6.6061 218 1.1331 0.4143 1.1331 1.0645
No log 6.6667 220 1.0733 0.4203 1.0733 1.0360
No log 6.7273 222 1.0011 0.4253 1.0011 1.0005
No log 6.7879 224 0.9967 0.4170 0.9967 0.9983
No log 6.8485 226 0.9830 0.4285 0.9830 0.9914
No log 6.9091 228 0.9576 0.4285 0.9576 0.9786
No log 6.9697 230 0.9811 0.4285 0.9811 0.9905
No log 7.0303 232 1.0022 0.3751 1.0022 1.0011
No log 7.0909 234 1.0155 0.3761 1.0155 1.0077
No log 7.1515 236 0.9988 0.3827 0.9988 0.9994
No log 7.2121 238 0.9728 0.3882 0.9728 0.9863
No log 7.2727 240 0.9222 0.4130 0.9222 0.9603
No log 7.3333 242 0.8832 0.4310 0.8832 0.9398
No log 7.3939 244 0.8446 0.4336 0.8446 0.9190
No log 7.4545 246 0.8468 0.4336 0.8468 0.9202
No log 7.5152 248 0.8930 0.4253 0.8930 0.9450
No log 7.5758 250 0.9583 0.3765 0.9583 0.9789
No log 7.6364 252 1.0547 0.3835 1.0547 1.0270
No log 7.6970 254 1.1004 0.3977 1.1004 1.0490
No log 7.7576 256 1.1197 0.3977 1.1197 1.0582
No log 7.8182 258 1.1149 0.3868 1.1149 1.0559
No log 7.8788 260 1.0624 0.3752 1.0624 1.0307
No log 7.9394 262 0.9899 0.4100 0.9899 0.9949
No log 8.0 264 0.9701 0.4100 0.9701 0.9849
No log 8.0606 266 0.9992 0.4031 0.9992 0.9996
No log 8.1212 268 1.0640 0.3905 1.0640 1.0315
No log 8.1818 270 1.1065 0.3956 1.1065 1.0519
No log 8.2424 272 1.1067 0.3705 1.1067 1.0520
No log 8.3030 274 1.0642 0.3718 1.0642 1.0316
No log 8.3636 276 0.9995 0.3696 0.9995 0.9998
No log 8.4242 278 0.9441 0.3925 0.9441 0.9716
No log 8.4848 280 0.9349 0.3944 0.9349 0.9669
No log 8.5455 282 0.9575 0.3833 0.9575 0.9785
No log 8.6061 284 0.9865 0.3647 0.9865 0.9932
No log 8.6667 286 1.0193 0.3774 1.0193 1.0096
No log 8.7273 288 1.0661 0.3707 1.0661 1.0325
No log 8.7879 290 1.0859 0.3598 1.0859 1.0421
No log 8.8485 292 1.0699 0.3598 1.0699 1.0344
No log 8.9091 294 1.0408 0.3774 1.0408 1.0202
No log 8.9697 296 1.0261 0.3774 1.0261 1.0130
No log 9.0303 298 1.0319 0.3827 1.0319 1.0158
No log 9.0909 300 1.0249 0.3827 1.0249 1.0124
No log 9.1515 302 1.0162 0.3827 1.0162 1.0081
No log 9.2121 304 1.0095 0.4081 1.0095 1.0048
No log 9.2727 306 1.0031 0.4081 1.0031 1.0015
No log 9.3333 308 0.9966 0.3867 0.9966 0.9983
No log 9.3939 310 0.9899 0.3818 0.9899 0.9949
No log 9.4545 312 0.9955 0.3867 0.9955 0.9977
No log 9.5152 314 1.0056 0.3935 1.0056 1.0028
No log 9.5758 316 1.0007 0.3935 1.0007 1.0003
No log 9.6364 318 0.9925 0.3735 0.9925 0.9962
No log 9.6970 320 0.9888 0.3735 0.9888 0.9944
No log 9.7576 322 0.9880 0.3735 0.9880 0.9940
No log 9.8182 324 0.9834 0.3735 0.9834 0.9916
No log 9.8788 326 0.9818 0.3696 0.9818 0.9908
No log 9.9394 328 0.9798 0.3696 0.9798 0.9899
No log 10.0 330 0.9791 0.3696 0.9791 0.9895

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k6_task2_organization

Finetuned
(4019)
this model