ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k6_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0521
  • Qwk: 0.6356
  • Mse: 1.0521
  • Rmse: 1.0257

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0769 2 2.2971 0.0408 2.2971 1.5156
No log 0.1538 4 1.5140 0.1452 1.5140 1.2305
No log 0.2308 6 1.5424 0.1127 1.5424 1.2419
No log 0.3077 8 1.5203 0.1209 1.5203 1.2330
No log 0.3846 10 1.4126 0.1209 1.4126 1.1885
No log 0.4615 12 1.3358 0.2146 1.3358 1.1557
No log 0.5385 14 1.2463 0.1888 1.2463 1.1164
No log 0.6154 16 1.2355 0.1833 1.2355 1.1115
No log 0.6923 18 1.2319 0.1833 1.2319 1.1099
No log 0.7692 20 1.2245 0.2058 1.2245 1.1066
No log 0.8462 22 1.2965 0.2973 1.2965 1.1386
No log 0.9231 24 1.4726 0.3800 1.4726 1.2135
No log 1.0 26 1.4855 0.4 1.4855 1.2188
No log 1.0769 28 1.3656 0.4010 1.3656 1.1686
No log 1.1538 30 1.3523 0.4422 1.3523 1.1629
No log 1.2308 32 1.3865 0.4207 1.3865 1.1775
No log 1.3077 34 1.3092 0.3916 1.3092 1.1442
No log 1.3846 36 1.2636 0.3594 1.2636 1.1241
No log 1.4615 38 1.1814 0.3233 1.1814 1.0869
No log 1.5385 40 1.1816 0.4005 1.1816 1.0870
No log 1.6154 42 1.2862 0.3858 1.2862 1.1341
No log 1.6923 44 1.3741 0.3412 1.3741 1.1722
No log 1.7692 46 1.4462 0.3636 1.4462 1.2026
No log 1.8462 48 1.3865 0.4111 1.3865 1.1775
No log 1.9231 50 1.2367 0.4330 1.2367 1.1120
No log 2.0 52 1.1138 0.4474 1.1138 1.0554
No log 2.0769 54 1.1193 0.4523 1.1193 1.0580
No log 2.1538 56 1.1506 0.4746 1.1506 1.0727
No log 2.2308 58 1.0693 0.4729 1.0693 1.0341
No log 2.3077 60 1.0314 0.5097 1.0314 1.0156
No log 2.3846 62 1.0508 0.5204 1.0508 1.0251
No log 2.4615 64 1.1511 0.5275 1.1511 1.0729
No log 2.5385 66 1.2534 0.5287 1.2534 1.1196
No log 2.6154 68 1.3037 0.5159 1.3037 1.1418
No log 2.6923 70 1.2760 0.5206 1.2760 1.1296
No log 2.7692 72 1.1707 0.5215 1.1707 1.0820
No log 2.8462 74 1.0895 0.5673 1.0895 1.0438
No log 2.9231 76 0.9993 0.6029 0.9993 0.9996
No log 3.0 78 1.0203 0.6101 1.0203 1.0101
No log 3.0769 80 1.1334 0.5852 1.1334 1.0646
No log 3.1538 82 1.2513 0.5704 1.2513 1.1186
No log 3.2308 84 1.1587 0.5548 1.1587 1.0764
No log 3.3077 86 1.0040 0.5576 1.0040 1.0020
No log 3.3846 88 0.9562 0.6123 0.9562 0.9779
No log 3.4615 90 1.0602 0.6103 1.0602 1.0296
No log 3.5385 92 1.1715 0.5781 1.1715 1.0824
No log 3.6154 94 1.4645 0.5156 1.4645 1.2102
No log 3.6923 96 1.5085 0.5415 1.5085 1.2282
No log 3.7692 98 1.2648 0.5853 1.2648 1.1246
No log 3.8462 100 1.0153 0.6190 1.0153 1.0076
No log 3.9231 102 1.0088 0.6306 1.0088 1.0044
No log 4.0 104 1.2046 0.5846 1.2046 1.0975
No log 4.0769 106 1.2576 0.5536 1.2576 1.1214
No log 4.1538 108 1.1769 0.5935 1.1769 1.0848
No log 4.2308 110 1.1531 0.5901 1.1531 1.0738
No log 4.3077 112 1.0363 0.5928 1.0363 1.0180
No log 4.3846 114 0.9253 0.6155 0.9253 0.9619
No log 4.4615 116 0.8575 0.6041 0.8575 0.9260
No log 4.5385 118 0.9263 0.6185 0.9263 0.9624
No log 4.6154 120 1.1470 0.6286 1.1470 1.0710
No log 4.6923 122 1.3521 0.5688 1.3521 1.1628
No log 4.7692 124 1.3559 0.5726 1.3559 1.1644
No log 4.8462 126 1.2136 0.5971 1.2136 1.1016
No log 4.9231 128 1.0325 0.6479 1.0325 1.0161
No log 5.0 130 0.9515 0.6442 0.9515 0.9755
No log 5.0769 132 1.0295 0.6495 1.0295 1.0146
No log 5.1538 134 1.2658 0.5963 1.2658 1.1251
No log 5.2308 136 1.5924 0.5231 1.5924 1.2619
No log 5.3077 138 1.6170 0.5142 1.6170 1.2716
No log 5.3846 140 1.4158 0.5573 1.4158 1.1899
No log 5.4615 142 1.1395 0.6194 1.1395 1.0675
No log 5.5385 144 0.9004 0.6205 0.9004 0.9489
No log 5.6154 146 0.8289 0.6305 0.8289 0.9104
No log 5.6923 148 0.8472 0.6167 0.8472 0.9204
No log 5.7692 150 0.9623 0.6085 0.9623 0.9810
No log 5.8462 152 1.1396 0.6296 1.1396 1.0675
No log 5.9231 154 1.2739 0.5901 1.2739 1.1287
No log 6.0 156 1.2963 0.5706 1.2963 1.1385
No log 6.0769 158 1.1838 0.5907 1.1838 1.0880
No log 6.1538 160 1.1119 0.5911 1.1119 1.0545
No log 6.2308 162 1.1158 0.5930 1.1158 1.0563
No log 6.3077 164 1.0863 0.6092 1.0863 1.0423
No log 6.3846 166 1.0388 0.6357 1.0388 1.0192
No log 6.4615 168 1.0300 0.6357 1.0300 1.0149
No log 6.5385 170 1.0978 0.6252 1.0978 1.0478
No log 6.6154 172 1.2332 0.5912 1.2332 1.1105
No log 6.6923 174 1.3176 0.5696 1.3176 1.1479
No log 6.7692 176 1.2786 0.5797 1.2786 1.1307
No log 6.8462 178 1.1326 0.5928 1.1326 1.0643
No log 6.9231 180 0.9465 0.6607 0.9465 0.9729
No log 7.0 182 0.8592 0.6742 0.8592 0.9269
No log 7.0769 184 0.8576 0.6877 0.8576 0.9261
No log 7.1538 186 0.9054 0.6558 0.9054 0.9515
No log 7.2308 188 0.9988 0.6465 0.9988 0.9994
No log 7.3077 190 1.0783 0.6341 1.0783 1.0384
No log 7.3846 192 1.1416 0.5997 1.1416 1.0685
No log 7.4615 194 1.1384 0.5819 1.1384 1.0670
No log 7.5385 196 1.1502 0.5799 1.1502 1.0725
No log 7.6154 198 1.1835 0.5858 1.1835 1.0879
No log 7.6923 200 1.2209 0.5786 1.2209 1.1049
No log 7.7692 202 1.2261 0.5844 1.2261 1.1073
No log 7.8462 204 1.1935 0.5825 1.1935 1.0925
No log 7.9231 206 1.1141 0.5839 1.1141 1.0555
No log 8.0 208 1.0007 0.6335 1.0007 1.0003
No log 8.0769 210 0.9092 0.6911 0.9092 0.9535
No log 8.1538 212 0.8981 0.6911 0.8981 0.9477
No log 8.2308 214 0.9408 0.6962 0.9408 0.9699
No log 8.3077 216 1.0127 0.6457 1.0127 1.0064
No log 8.3846 218 1.0825 0.6340 1.0825 1.0404
No log 8.4615 220 1.1155 0.6195 1.1155 1.0562
No log 8.5385 222 1.1755 0.6008 1.1755 1.0842
No log 8.6154 224 1.2009 0.5816 1.2009 1.0959
No log 8.6923 226 1.1795 0.5906 1.1795 1.0860
No log 8.7692 228 1.1215 0.6040 1.1215 1.0590
No log 8.8462 230 1.0641 0.6176 1.0641 1.0316
No log 8.9231 232 1.0391 0.6334 1.0391 1.0193
No log 9.0 234 1.0428 0.6334 1.0428 1.0212
No log 9.0769 236 1.0679 0.6234 1.0679 1.0334
No log 9.1538 238 1.0750 0.6395 1.0750 1.0368
No log 9.2308 240 1.0653 0.6356 1.0653 1.0321
No log 9.3077 242 1.0433 0.6294 1.0433 1.0214
No log 9.3846 244 1.0228 0.6294 1.0228 1.0113
No log 9.4615 246 1.0139 0.6357 1.0139 1.0069
No log 9.5385 248 1.0160 0.6476 1.0160 1.0080
No log 9.6154 250 1.0258 0.6419 1.0258 1.0128
No log 9.6923 252 1.0358 0.6419 1.0358 1.0178
No log 9.7692 254 1.0434 0.6356 1.0434 1.0214
No log 9.8462 256 1.0460 0.6356 1.0460 1.0227
No log 9.9231 258 1.0499 0.6356 1.0499 1.0246
No log 10.0 260 1.0521 0.6356 1.0521 1.0257

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k6_task5_organization

Finetuned
(4023)
this model