ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k3_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5863
  • Qwk: 0.7383
  • Mse: 0.5863
  • Rmse: 0.7657

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 2.1424 0.0392 2.1424 1.4637
No log 0.3077 4 1.4228 0.2019 1.4228 1.1928
No log 0.4615 6 1.3512 0.2156 1.3512 1.1624
No log 0.6154 8 1.5465 0.3246 1.5465 1.2436
No log 0.7692 10 1.4248 0.1797 1.4248 1.1937
No log 0.9231 12 1.4367 0.1549 1.4367 1.1986
No log 1.0769 14 1.4150 0.1639 1.4150 1.1895
No log 1.2308 16 1.5234 0.3116 1.5234 1.2343
No log 1.3846 18 1.3895 0.3454 1.3895 1.1788
No log 1.5385 20 1.2064 0.3051 1.2064 1.0984
No log 1.6923 22 1.1051 0.3611 1.1051 1.0512
No log 1.8462 24 1.0332 0.4010 1.0332 1.0165
No log 2.0 26 0.9844 0.4430 0.9844 0.9922
No log 2.1538 28 0.9614 0.5330 0.9614 0.9805
No log 2.3077 30 0.9503 0.5319 0.9503 0.9749
No log 2.4615 32 0.8952 0.5312 0.8952 0.9462
No log 2.6154 34 0.8392 0.6225 0.8392 0.9161
No log 2.7692 36 0.8736 0.5780 0.8736 0.9346
No log 2.9231 38 0.8712 0.5494 0.8712 0.9334
No log 3.0769 40 0.9476 0.4966 0.9476 0.9734
No log 3.2308 42 0.8074 0.6178 0.8074 0.8985
No log 3.3846 44 0.7086 0.6766 0.7086 0.8418
No log 3.5385 46 0.6797 0.6711 0.6797 0.8244
No log 3.6923 48 0.6767 0.6776 0.6767 0.8226
No log 3.8462 50 0.6620 0.6891 0.6620 0.8136
No log 4.0 52 0.6544 0.6882 0.6544 0.8090
No log 4.1538 54 0.6656 0.6988 0.6656 0.8158
No log 4.3077 56 0.6465 0.7035 0.6465 0.8041
No log 4.4615 58 0.6427 0.7130 0.6427 0.8017
No log 4.6154 60 0.6437 0.6957 0.6437 0.8023
No log 4.7692 62 0.6245 0.7088 0.6245 0.7902
No log 4.9231 64 0.6212 0.7432 0.6212 0.7882
No log 5.0769 66 0.6527 0.7450 0.6527 0.8079
No log 5.2308 68 0.6142 0.7336 0.6142 0.7837
No log 5.3846 70 0.6316 0.7206 0.6316 0.7947
No log 5.5385 72 0.6868 0.6522 0.6868 0.8287
No log 5.6923 74 0.6916 0.6522 0.6916 0.8316
No log 5.8462 76 0.6466 0.7025 0.6466 0.8041
No log 6.0 78 0.6069 0.7184 0.6069 0.7791
No log 6.1538 80 0.6068 0.7184 0.6068 0.7790
No log 6.3077 82 0.6161 0.7123 0.6161 0.7849
No log 6.4615 84 0.6503 0.6501 0.6503 0.8064
No log 6.6154 86 0.6488 0.6501 0.6488 0.8055
No log 6.7692 88 0.6133 0.7322 0.6133 0.7831
No log 6.9231 90 0.6196 0.7493 0.6196 0.7871
No log 7.0769 92 0.6246 0.7469 0.6246 0.7903
No log 7.2308 94 0.6128 0.7482 0.6128 0.7828
No log 7.3846 96 0.5996 0.7523 0.5996 0.7743
No log 7.5385 98 0.6055 0.7370 0.6055 0.7781
No log 7.6923 100 0.6069 0.7416 0.6069 0.7791
No log 7.8462 102 0.5985 0.7344 0.5985 0.7736
No log 8.0 104 0.5919 0.7383 0.5919 0.7693
No log 8.1538 106 0.5890 0.7375 0.5890 0.7674
No log 8.3077 108 0.5907 0.7474 0.5907 0.7685
No log 8.4615 110 0.5903 0.7554 0.5903 0.7683
No log 8.6154 112 0.5882 0.7554 0.5882 0.7670
No log 8.7692 114 0.5852 0.7251 0.5852 0.7650
No log 8.9231 116 0.5891 0.7407 0.5891 0.7675
No log 9.0769 118 0.5915 0.7399 0.5915 0.7691
No log 9.2308 120 0.5964 0.7272 0.5964 0.7723
No log 9.3846 122 0.5958 0.7360 0.5958 0.7719
No log 9.5385 124 0.5922 0.7399 0.5922 0.7696
No log 9.6923 126 0.5886 0.7367 0.5886 0.7672
No log 9.8462 128 0.5865 0.7383 0.5865 0.7659
No log 10.0 130 0.5863 0.7383 0.5863 0.7657

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k3_task5_organization

Finetuned
(4023)
this model