ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k4_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6238
  • Qwk: 0.3143
  • Mse: 0.6238
  • Rmse: 0.7898

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0870 2 3.1862 -0.0149 3.1862 1.7850
No log 0.1739 4 1.5063 -0.0070 1.5063 1.2273
No log 0.2609 6 0.9292 0.0154 0.9292 0.9640
No log 0.3478 8 0.5718 0.0569 0.5718 0.7562
No log 0.4348 10 0.6197 -0.0159 0.6197 0.7872
No log 0.5217 12 0.6197 0.0 0.6197 0.7872
No log 0.6087 14 0.5831 0.0 0.5831 0.7636
No log 0.6957 16 0.6344 0.1008 0.6344 0.7965
No log 0.7826 18 0.7904 0.1392 0.7904 0.8891
No log 0.8696 20 0.9634 0.0078 0.9634 0.9815
No log 0.9565 22 0.9495 0.0431 0.9495 0.9744
No log 1.0435 24 0.6932 0.1038 0.6932 0.8326
No log 1.1304 26 0.8008 0.0 0.8008 0.8949
No log 1.2174 28 1.5664 0.0345 1.5664 1.2516
No log 1.3043 30 1.2554 0.0118 1.2554 1.1204
No log 1.3913 32 0.6798 -0.0963 0.6798 0.8245
No log 1.4783 34 0.7198 0.0 0.7198 0.8484
No log 1.5652 36 0.5975 0.0 0.5975 0.7730
No log 1.6522 38 0.6422 0.1304 0.6422 0.8013
No log 1.7391 40 0.7374 0.1515 0.7374 0.8587
No log 1.8261 42 0.7895 0.1930 0.7895 0.8886
No log 1.9130 44 0.7567 0.0857 0.7567 0.8699
No log 2.0 46 0.6101 0.1020 0.6101 0.7811
No log 2.0870 48 0.5858 0.0 0.5858 0.7653
No log 2.1739 50 0.5961 0.0 0.5961 0.7721
No log 2.2609 52 0.5845 0.0909 0.5845 0.7645
No log 2.3478 54 0.7893 0.1351 0.7893 0.8884
No log 2.4348 56 1.0004 0.1220 1.0004 1.0002
No log 2.5217 58 0.9796 0.1220 0.9796 0.9897
No log 2.6087 60 0.8244 0.1644 0.8244 0.9080
No log 2.6957 62 0.6330 0.1186 0.6330 0.7956
No log 2.7826 64 0.5820 -0.0196 0.5820 0.7629
No log 2.8696 66 0.5777 0.1888 0.5777 0.7601
No log 2.9565 68 0.6395 0.2970 0.6395 0.7997
No log 3.0435 70 0.5984 0.2516 0.5984 0.7736
No log 3.1304 72 0.5057 0.1892 0.5057 0.7111
No log 3.2174 74 0.6451 0.2157 0.6451 0.8032
No log 3.3043 76 0.6688 0.2077 0.6688 0.8178
No log 3.3913 78 0.5754 0.2000 0.5754 0.7586
No log 3.4783 80 0.5170 0.1788 0.5170 0.7190
No log 3.5652 82 0.5106 0.2704 0.5106 0.7146
No log 3.6522 84 0.5492 0.3073 0.5492 0.7411
No log 3.7391 86 0.5183 0.3333 0.5183 0.7199
No log 3.8261 88 0.6397 0.2079 0.6397 0.7998
No log 3.9130 90 1.0533 0.2065 1.0533 1.0263
No log 4.0 92 1.0456 0.2131 1.0456 1.0226
No log 4.0870 94 0.7287 0.2442 0.7287 0.8536
No log 4.1739 96 0.6168 0.2549 0.6168 0.7853
No log 4.2609 98 0.7530 0.2227 0.7530 0.8678
No log 4.3478 100 0.8038 0.2838 0.8038 0.8965
No log 4.4348 102 0.6810 0.2161 0.6810 0.8252
No log 4.5217 104 0.6466 0.1186 0.6466 0.8041
No log 4.6087 106 0.7086 0.1600 0.7086 0.8418
No log 4.6957 108 0.6853 0.1600 0.6853 0.8278
No log 4.7826 110 0.5760 0.1086 0.5760 0.7589
No log 4.8696 112 0.6470 0.2941 0.6470 0.8044
No log 4.9565 114 0.7274 0.2821 0.7274 0.8529
No log 5.0435 116 0.8475 0.2829 0.8475 0.9206
No log 5.1304 118 0.7097 0.2775 0.7097 0.8425
No log 5.2174 120 0.6660 0.2982 0.6660 0.8161
No log 5.3043 122 0.6022 0.4123 0.6022 0.7760
No log 5.3913 124 0.6507 0.4383 0.6507 0.8067
No log 5.4783 126 0.6149 0.3982 0.6149 0.7842
No log 5.5652 128 0.5994 0.4027 0.5994 0.7742
No log 5.6522 130 0.5849 0.3973 0.5849 0.7648
No log 5.7391 132 0.6693 0.4236 0.6693 0.8181
No log 5.8261 134 0.8193 0.2640 0.8193 0.9051
No log 5.9130 136 0.8652 0.2389 0.8652 0.9302
No log 6.0 138 0.8833 0.2389 0.8833 0.9398
No log 6.0870 140 0.7653 0.3333 0.7653 0.8748
No log 6.1739 142 0.6552 0.3091 0.6552 0.8094
No log 6.2609 144 0.6699 0.2961 0.6699 0.8185
No log 6.3478 146 0.6794 0.3306 0.6794 0.8242
No log 6.4348 148 0.8103 0.2961 0.8103 0.9002
No log 6.5217 150 1.0933 0.1822 1.0933 1.0456
No log 6.6087 152 1.1452 0.1367 1.1452 1.0702
No log 6.6957 154 0.9101 0.1803 0.9101 0.9540
No log 6.7826 156 0.6609 0.3333 0.6609 0.8130
No log 6.8696 158 0.6195 0.3427 0.6195 0.7871
No log 6.9565 160 0.6210 0.3365 0.6210 0.7881
No log 7.0435 162 0.6222 0.3645 0.6222 0.7888
No log 7.1304 164 0.6285 0.3722 0.6285 0.7928
No log 7.2174 166 0.7106 0.2982 0.7106 0.8430
No log 7.3043 168 0.8324 0.3188 0.8324 0.9124
No log 7.3913 170 0.8147 0.2775 0.8147 0.9026
No log 7.4783 172 0.7574 0.2356 0.7574 0.8703
No log 7.5652 174 0.6774 0.3665 0.6774 0.8230
No log 7.6522 176 0.6319 0.3455 0.6319 0.7949
No log 7.7391 178 0.6340 0.3739 0.6340 0.7962
No log 7.8261 180 0.6514 0.3886 0.6514 0.8071
No log 7.9130 182 0.7147 0.3684 0.7147 0.8454
No log 8.0 184 0.7701 0.2320 0.7701 0.8776
No log 8.0870 186 0.7174 0.3613 0.7174 0.8470
No log 8.1739 188 0.6104 0.4188 0.6104 0.7813
No log 8.2609 190 0.5496 0.3665 0.5496 0.7414
No log 8.3478 192 0.5452 0.4010 0.5452 0.7384
No log 8.4348 194 0.5427 0.3962 0.5427 0.7367
No log 8.5217 196 0.5535 0.3761 0.5535 0.7440
No log 8.6087 198 0.6090 0.4188 0.6090 0.7804
No log 8.6957 200 0.6788 0.3648 0.6788 0.8239
No log 8.7826 202 0.7177 0.3580 0.7177 0.8472
No log 8.8696 204 0.6920 0.3277 0.6920 0.8319
No log 8.9565 206 0.6380 0.3939 0.6380 0.7988
No log 9.0435 208 0.5787 0.3684 0.5787 0.7607
No log 9.1304 210 0.5580 0.3684 0.5580 0.7470
No log 9.2174 212 0.5506 0.3585 0.5506 0.7420
No log 9.3043 214 0.5491 0.3585 0.5491 0.7410
No log 9.3913 216 0.5550 0.3684 0.5550 0.7450
No log 9.4783 218 0.5647 0.3684 0.5647 0.7515
No log 9.5652 220 0.5790 0.3684 0.5790 0.7609
No log 9.6522 222 0.5947 0.3398 0.5947 0.7711
No log 9.7391 224 0.6088 0.3427 0.6088 0.7803
No log 9.8261 226 0.6191 0.3427 0.6191 0.7868
No log 9.9130 228 0.6229 0.3143 0.6229 0.7893
No log 10.0 230 0.6238 0.3143 0.6238 0.7898

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k4_task3_organization

Finetuned
(4023)
this model