ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k5_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0447
  • Qwk: 0.6514
  • Mse: 1.0447
  • Rmse: 1.0221

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0909 2 2.3232 0.0408 2.3232 1.5242
No log 0.1818 4 1.4648 0.2161 1.4648 1.2103
No log 0.2727 6 1.4348 0.2084 1.4348 1.1978
No log 0.3636 8 1.7036 0.1767 1.7036 1.3052
No log 0.4545 10 1.8246 0.2006 1.8246 1.3508
No log 0.5455 12 1.8471 0.1664 1.8471 1.3591
No log 0.6364 14 1.7442 0.1563 1.7442 1.3207
No log 0.7273 16 1.6507 0.2426 1.6507 1.2848
No log 0.8182 18 1.7352 0.3125 1.7352 1.3173
No log 0.9091 20 1.5902 0.3182 1.5902 1.2610
No log 1.0 22 1.3922 0.3117 1.3922 1.1799
No log 1.0909 24 1.7006 0.0258 1.7006 1.3041
No log 1.1818 26 1.5474 0.1010 1.5474 1.2439
No log 1.2727 28 1.3040 0.3045 1.3040 1.1419
No log 1.3636 30 1.3249 0.1700 1.3249 1.1510
No log 1.4545 32 1.6112 0.3128 1.6112 1.2693
No log 1.5455 34 1.9729 0.2966 1.9729 1.4046
No log 1.6364 36 1.9542 0.2120 1.9542 1.3979
No log 1.7273 38 1.7118 0.1717 1.7118 1.3084
No log 1.8182 40 1.5914 0.1601 1.5914 1.2615
No log 1.9091 42 1.4697 0.2231 1.4697 1.2123
No log 2.0 44 1.3368 0.2096 1.3368 1.1562
No log 2.0909 46 1.2636 0.2485 1.2636 1.1241
No log 2.1818 48 1.2227 0.2558 1.2227 1.1057
No log 2.2727 50 1.2366 0.3142 1.2366 1.1120
No log 2.3636 52 1.4420 0.3763 1.4420 1.2008
No log 2.4545 54 1.7836 0.3649 1.7836 1.3355
No log 2.5455 56 1.9316 0.3566 1.9316 1.3898
No log 2.6364 58 1.7657 0.3821 1.7657 1.3288
No log 2.7273 60 1.4157 0.4308 1.4157 1.1898
No log 2.8182 62 1.1394 0.4401 1.1394 1.0674
No log 2.9091 64 1.0269 0.4775 1.0269 1.0134
No log 3.0 66 1.0422 0.4898 1.0422 1.0209
No log 3.0909 68 1.1028 0.4907 1.1028 1.0502
No log 3.1818 70 1.3033 0.4849 1.3033 1.1416
No log 3.2727 72 1.3677 0.5069 1.3677 1.1695
No log 3.3636 74 1.2803 0.5095 1.2803 1.1315
No log 3.4545 76 1.1342 0.5169 1.1342 1.0650
No log 3.5455 78 1.0745 0.5531 1.0745 1.0366
No log 3.6364 80 1.0209 0.5417 1.0209 1.0104
No log 3.7273 82 1.0009 0.5668 1.0009 1.0005
No log 3.8182 84 0.9177 0.5729 0.9177 0.9579
No log 3.9091 86 0.9219 0.5841 0.9219 0.9602
No log 4.0 88 0.9956 0.6172 0.9956 0.9978
No log 4.0909 90 1.2102 0.5747 1.2102 1.1001
No log 4.1818 92 1.5193 0.4991 1.5193 1.2326
No log 4.2727 94 1.7517 0.4799 1.7517 1.3235
No log 4.3636 96 1.7893 0.4721 1.7893 1.3377
No log 4.4545 98 1.5128 0.4981 1.5128 1.2300
No log 4.5455 100 1.3053 0.5122 1.3053 1.1425
No log 4.6364 102 1.2647 0.5287 1.2647 1.1246
No log 4.7273 104 1.3277 0.5305 1.3277 1.1522
No log 4.8182 106 1.2296 0.5536 1.2296 1.1089
No log 4.9091 108 1.1758 0.5866 1.1758 1.0844
No log 5.0 110 1.0572 0.6282 1.0572 1.0282
No log 5.0909 112 1.0511 0.6142 1.0511 1.0252
No log 5.1818 114 1.1897 0.5862 1.1897 1.0907
No log 5.2727 116 1.2201 0.5666 1.2201 1.1046
No log 5.3636 118 1.1365 0.5748 1.1365 1.0661
No log 5.4545 120 1.0203 0.6001 1.0203 1.0101
No log 5.5455 122 0.8857 0.5912 0.8857 0.9411
No log 5.6364 124 0.8467 0.6065 0.8467 0.9202
No log 5.7273 126 0.8237 0.6326 0.8237 0.9076
No log 5.8182 128 0.8497 0.6444 0.8497 0.9218
No log 5.9091 130 0.9360 0.6410 0.9360 0.9675
No log 6.0 132 1.0201 0.6565 1.0201 1.0100
No log 6.0909 134 1.0055 0.6565 1.0055 1.0027
No log 6.1818 136 1.0341 0.6542 1.0341 1.0169
No log 6.2727 138 1.1329 0.6118 1.1329 1.0644
No log 6.3636 140 1.1284 0.6078 1.1284 1.0622
No log 6.4545 142 1.0549 0.6482 1.0549 1.0271
No log 6.5455 144 0.9663 0.6790 0.9663 0.9830
No log 6.6364 146 0.9920 0.6597 0.9920 0.9960
No log 6.7273 148 1.0955 0.6201 1.0955 1.0466
No log 6.8182 150 1.1314 0.6097 1.1314 1.0637
No log 6.9091 152 1.1051 0.6010 1.1051 1.0512
No log 7.0 154 1.0924 0.5910 1.0924 1.0452
No log 7.0909 156 1.0017 0.64 1.0017 1.0008
No log 7.1818 158 0.9360 0.6556 0.9360 0.9674
No log 7.2727 160 0.9589 0.6416 0.9589 0.9792
No log 7.3636 162 0.9953 0.6507 0.9953 0.9976
No log 7.4545 164 1.0813 0.5984 1.0813 1.0398
No log 7.5455 166 1.2013 0.5620 1.2013 1.0960
No log 7.6364 168 1.2961 0.5546 1.2961 1.1385
No log 7.7273 170 1.3393 0.5508 1.3393 1.1573
No log 7.8182 172 1.2778 0.5495 1.2778 1.1304
No log 7.9091 174 1.2083 0.5864 1.2083 1.0992
No log 8.0 176 1.0975 0.6475 1.0975 1.0476
No log 8.0909 178 0.9997 0.6648 0.9997 0.9999
No log 8.1818 180 0.9557 0.6518 0.9557 0.9776
No log 8.2727 182 0.9609 0.6518 0.9609 0.9803
No log 8.3636 184 1.0188 0.6557 1.0188 1.0093
No log 8.4545 186 1.0847 0.6460 1.0847 1.0415
No log 8.5455 188 1.1420 0.6194 1.1420 1.0686
No log 8.6364 190 1.1294 0.6194 1.1294 1.0627
No log 8.7273 192 1.1020 0.6277 1.1020 1.0498
No log 8.8182 194 1.0859 0.6375 1.0859 1.0421
No log 8.9091 196 1.0673 0.6460 1.0673 1.0331
No log 9.0 198 1.0539 0.6475 1.0539 1.0266
No log 9.0909 200 1.0286 0.6739 1.0286 1.0142
No log 9.1818 202 1.0382 0.6739 1.0382 1.0189
No log 9.2727 204 1.0390 0.6739 1.0390 1.0193
No log 9.3636 206 1.0310 0.6852 1.0310 1.0154
No log 9.4545 208 1.0382 0.6602 1.0382 1.0189
No log 9.5455 210 1.0491 0.6514 1.0491 1.0242
No log 9.6364 212 1.0566 0.6475 1.0566 1.0279
No log 9.7273 214 1.0581 0.6475 1.0581 1.0286
No log 9.8182 216 1.0525 0.6475 1.0525 1.0259
No log 9.9091 218 1.0474 0.6514 1.0474 1.0234
No log 10.0 220 1.0447 0.6514 1.0447 1.0221

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k5_task5_organization

Finetuned
(4023)
this model