ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k4_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7446
  • Qwk: 0.2372
  • Mse: 0.7446
  • Rmse: 0.8629

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1 2 3.2078 -0.0028 3.2078 1.7910
No log 0.2 4 1.5451 0.0255 1.5451 1.2430
No log 0.3 6 1.0837 0.0335 1.0837 1.0410
No log 0.4 8 0.7208 0.1398 0.7208 0.8490
No log 0.5 10 0.5880 0.0909 0.5880 0.7668
No log 0.6 12 0.6068 0.0534 0.6068 0.7790
No log 0.7 14 0.6055 0.1278 0.6055 0.7782
No log 0.8 16 0.5965 0.0222 0.5965 0.7723
No log 0.9 18 0.6167 0.0897 0.6167 0.7853
No log 1.0 20 0.8301 0.0476 0.8301 0.9111
No log 1.1 22 0.7052 0.0429 0.7052 0.8398
No log 1.2 24 0.9406 0.0857 0.9406 0.9699
No log 1.3 26 0.9606 0.1333 0.9606 0.9801
No log 1.4 28 0.8483 0.1030 0.8483 0.9210
No log 1.5 30 0.6657 0.0625 0.6657 0.8159
No log 1.6 32 0.7859 0.1765 0.7859 0.8865
No log 1.7 34 1.5105 0.0833 1.5105 1.2290
No log 1.8 36 1.3564 0.0833 1.3564 1.1646
No log 1.9 38 0.7619 0.1919 0.7619 0.8729
No log 2.0 40 0.5993 0.0365 0.5993 0.7741
No log 2.1 42 0.6835 0.0638 0.6835 0.8268
No log 2.2 44 0.7053 0.1233 0.7053 0.8398
No log 2.3 46 0.6824 0.1667 0.6824 0.8261
No log 2.4 48 0.6650 0.0452 0.6650 0.8155
No log 2.5 50 0.7971 0.1414 0.7971 0.8928
No log 2.6 52 0.6954 0.1628 0.6954 0.8339
No log 2.7 54 0.6811 0.1565 0.6811 0.8253
No log 2.8 56 0.6948 0.1467 0.6948 0.8336
No log 2.9 58 0.7502 0.2169 0.7502 0.8662
No log 3.0 60 0.6620 0.1373 0.6620 0.8136
No log 3.1 62 0.6828 0.1351 0.6828 0.8263
No log 3.2 64 0.7646 0.0899 0.7646 0.8744
No log 3.3 66 0.8254 0.0531 0.8254 0.9085
No log 3.4 68 0.8442 0.0495 0.8442 0.9188
No log 3.5 70 0.8739 0.0359 0.8739 0.9348
No log 3.6 72 0.9017 0.0189 0.9017 0.9496
No log 3.7 74 0.9267 0.1131 0.9267 0.9627
No log 3.8 76 1.0730 0.0367 1.0730 1.0359
No log 3.9 78 1.6611 0.0327 1.6611 1.2889
No log 4.0 80 1.7461 0.0096 1.7461 1.3214
No log 4.1 82 1.1546 0.1161 1.1546 1.0745
No log 4.2 84 0.8773 0.0493 0.8773 0.9366
No log 4.3 86 0.8552 0.0707 0.8552 0.9248
No log 4.4 88 1.1661 0.0938 1.1661 1.0799
No log 4.5 90 1.6476 0.0809 1.6476 1.2836
No log 4.6 92 1.4167 0.0704 1.4167 1.1902
No log 4.7 94 0.8427 0.1304 0.8427 0.9180
No log 4.8 96 0.7782 0.0980 0.7782 0.8822
No log 4.9 98 0.7502 0.1230 0.7502 0.8662
No log 5.0 100 0.8006 0.2086 0.8006 0.8948
No log 5.1 102 1.2549 0.0365 1.2549 1.1202
No log 5.2 104 1.6616 0.0881 1.6616 1.2890
No log 5.3 106 1.5159 0.1068 1.5159 1.2312
No log 5.4 108 0.9729 0.1790 0.9729 0.9863
No log 5.5 110 0.7998 0.2850 0.7998 0.8943
No log 5.6 112 0.7470 0.2811 0.7470 0.8643
No log 5.7 114 0.7517 0.2593 0.7517 0.8670
No log 5.8 116 0.8336 0.1790 0.8336 0.9130
No log 5.9 118 0.8887 0.1718 0.8887 0.9427
No log 6.0 120 0.8926 0.2356 0.8926 0.9448
No log 6.1 122 0.9367 0.1351 0.9367 0.9679
No log 6.2 124 0.9827 0.1366 0.9827 0.9913
No log 6.3 126 0.9026 0.1579 0.9026 0.9500
No log 6.4 128 0.8715 0.2667 0.8715 0.9335
No log 6.5 130 0.9417 0.1652 0.9417 0.9704
No log 6.6 132 0.9381 0.1652 0.9381 0.9685
No log 6.7 134 0.9616 0.1074 0.9616 0.9806
No log 6.8 136 0.8501 0.2320 0.8501 0.9220
No log 6.9 138 0.7571 0.3004 0.7571 0.8701
No log 7.0 140 0.7832 0.2775 0.7832 0.8850
No log 7.1 142 0.8966 0.1441 0.8966 0.9469
No log 7.2 144 0.8297 0.1652 0.8297 0.9109
No log 7.3 146 0.7365 0.3455 0.7365 0.8582
No log 7.4 148 0.7629 0.2857 0.7629 0.8734
No log 7.5 150 0.8672 0.2000 0.8672 0.9313
No log 7.6 152 0.8516 0.1660 0.8516 0.9228
No log 7.7 154 0.8877 0.1724 0.8877 0.9422
No log 7.8 156 0.8570 0.1724 0.8570 0.9257
No log 7.9 158 0.7572 0.25 0.7572 0.8702
No log 8.0 160 0.7600 0.2579 0.7600 0.8718
No log 8.1 162 0.8577 0.1724 0.8577 0.9261
No log 8.2 164 1.0198 0.1746 1.0198 1.0098
No log 8.3 166 1.0129 0.1486 1.0129 1.0064
No log 8.4 168 0.8934 0.1441 0.8934 0.9452
No log 8.5 170 0.8042 0.2070 0.8042 0.8968
No log 8.6 172 0.7367 0.3846 0.7367 0.8583
No log 8.7 174 0.7252 0.3846 0.7252 0.8516
No log 8.8 176 0.7370 0.3171 0.7370 0.8585
No log 8.9 178 0.7664 0.1781 0.7664 0.8754
No log 9.0 180 0.7724 0.1781 0.7724 0.8789
No log 9.1 182 0.7666 0.1712 0.7666 0.8755
No log 9.2 184 0.7449 0.2762 0.7449 0.8631
No log 9.3 186 0.7196 0.4286 0.7196 0.8483
No log 9.4 188 0.7203 0.4286 0.7203 0.8487
No log 9.5 190 0.7269 0.4286 0.7269 0.8526
No log 9.6 192 0.7396 0.3427 0.7396 0.8600
No log 9.7 194 0.7372 0.3427 0.7372 0.8586
No log 9.8 196 0.7417 0.3427 0.7417 0.8612
No log 9.9 198 0.7445 0.2372 0.7445 0.8628
No log 10.0 200 0.7446 0.2372 0.7446 0.8629

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k4_task3_organization

Finetuned
(4023)
this model