ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k4_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6463
  • Qwk: 0.7500
  • Mse: 0.6463
  • Rmse: 0.8039

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1111 2 2.2588 -0.0233 2.2588 1.5029
No log 0.2222 4 1.5405 0.1363 1.5405 1.2412
No log 0.3333 6 1.3849 0.1718 1.3849 1.1768
No log 0.4444 8 1.3620 0.2065 1.3620 1.1671
No log 0.5556 10 1.1089 0.2871 1.1089 1.0530
No log 0.6667 12 1.0663 0.3276 1.0663 1.0326
No log 0.7778 14 1.1024 0.3826 1.1024 1.0500
No log 0.8889 16 0.9647 0.4285 0.9647 0.9822
No log 1.0 18 1.0225 0.5138 1.0225 1.0112
No log 1.1111 20 1.0196 0.5265 1.0196 1.0097
No log 1.2222 22 0.8975 0.4882 0.8975 0.9473
No log 1.3333 24 0.8989 0.4944 0.8989 0.9481
No log 1.4444 26 0.8669 0.5039 0.8669 0.9311
No log 1.5556 28 0.8673 0.5799 0.8673 0.9313
No log 1.6667 30 0.9829 0.5381 0.9829 0.9914
No log 1.7778 32 0.9476 0.5394 0.9476 0.9735
No log 1.8889 34 0.8114 0.6330 0.8114 0.9008
No log 2.0 36 0.7608 0.6596 0.7608 0.8722
No log 2.1111 38 0.7687 0.6763 0.7687 0.8768
No log 2.2222 40 0.7960 0.6752 0.7960 0.8922
No log 2.3333 42 0.9475 0.6040 0.9475 0.9734
No log 2.4444 44 1.0283 0.5675 1.0283 1.0140
No log 2.5556 46 0.8513 0.6492 0.8513 0.9226
No log 2.6667 48 0.7050 0.7126 0.7050 0.8396
No log 2.7778 50 0.7166 0.7160 0.7166 0.8465
No log 2.8889 52 0.7844 0.6501 0.7844 0.8857
No log 3.0 54 0.7967 0.6460 0.7967 0.8926
No log 3.1111 56 0.7384 0.6636 0.7384 0.8593
No log 3.2222 58 0.7140 0.6737 0.7140 0.8450
No log 3.3333 60 0.7171 0.6737 0.7171 0.8468
No log 3.4444 62 0.6771 0.6953 0.6771 0.8229
No log 3.5556 64 0.6525 0.7380 0.6525 0.8078
No log 3.6667 66 0.7088 0.7248 0.7088 0.8419
No log 3.7778 68 0.7669 0.7362 0.7669 0.8758
No log 3.8889 70 0.7363 0.7481 0.7363 0.8581
No log 4.0 72 0.7227 0.7539 0.7227 0.8501
No log 4.1111 74 0.7545 0.7266 0.7545 0.8686
No log 4.2222 76 0.7285 0.7396 0.7285 0.8535
No log 4.3333 78 0.6981 0.7559 0.6981 0.8355
No log 4.4444 80 0.6675 0.7575 0.6675 0.8170
No log 4.5556 82 0.6565 0.7575 0.6565 0.8103
No log 4.6667 84 0.6485 0.7603 0.6485 0.8053
No log 4.7778 86 0.6736 0.7575 0.6736 0.8207
No log 4.8889 88 0.6438 0.7651 0.6438 0.8023
No log 5.0 90 0.6325 0.7353 0.6325 0.7953
No log 5.1111 92 0.6373 0.7416 0.6373 0.7983
No log 5.2222 94 0.6278 0.7400 0.6278 0.7923
No log 5.3333 96 0.6253 0.7361 0.6253 0.7907
No log 5.4444 98 0.6378 0.7298 0.6378 0.7986
No log 5.5556 100 0.6253 0.7298 0.6253 0.7908
No log 5.6667 102 0.6554 0.7269 0.6554 0.8096
No log 5.7778 104 0.8158 0.7220 0.8158 0.9032
No log 5.8889 106 0.8528 0.7080 0.8528 0.9235
No log 6.0 108 0.7530 0.7315 0.7530 0.8678
No log 6.1111 110 0.6391 0.7369 0.6391 0.7994
No log 6.2222 112 0.6448 0.7054 0.6448 0.8030
No log 6.3333 114 0.6401 0.7054 0.6401 0.8001
No log 6.4444 116 0.6179 0.7321 0.6179 0.7861
No log 6.5556 118 0.6641 0.7501 0.6641 0.8149
No log 6.6667 120 0.7534 0.7467 0.7534 0.8680
No log 6.7778 122 0.7540 0.7392 0.7540 0.8684
No log 6.8889 124 0.6838 0.7482 0.6838 0.8269
No log 7.0 126 0.6095 0.7552 0.6095 0.7807
No log 7.1111 128 0.6003 0.7357 0.6003 0.7748
No log 7.2222 130 0.6214 0.7310 0.6214 0.7883
No log 7.3333 132 0.6161 0.7287 0.6161 0.7849
No log 7.4444 134 0.6003 0.7511 0.6003 0.7748
No log 7.5556 136 0.6050 0.7619 0.6050 0.7778
No log 7.6667 138 0.6248 0.7552 0.6248 0.7904
No log 7.7778 140 0.6403 0.7640 0.6403 0.8002
No log 7.8889 142 0.6684 0.7387 0.6684 0.8176
No log 8.0 144 0.6836 0.7442 0.6836 0.8268
No log 8.1111 146 0.6799 0.7442 0.6799 0.8245
No log 8.2222 148 0.6505 0.7659 0.6505 0.8065
No log 8.3333 150 0.6160 0.7619 0.6160 0.7849
No log 8.4444 152 0.6023 0.7563 0.6023 0.7761
No log 8.5556 154 0.6013 0.7587 0.6013 0.7755
No log 8.6667 156 0.6027 0.7587 0.6027 0.7763
No log 8.7778 158 0.6033 0.7587 0.6033 0.7767
No log 8.8889 160 0.6072 0.7563 0.6072 0.7792
No log 9.0 162 0.6137 0.7619 0.6137 0.7834
No log 9.1111 164 0.6178 0.7619 0.6178 0.7860
No log 9.2222 166 0.6243 0.7619 0.6243 0.7901
No log 9.3333 168 0.6272 0.7619 0.6272 0.7920
No log 9.4444 170 0.6276 0.7619 0.6276 0.7922
No log 9.5556 172 0.6313 0.7619 0.6313 0.7945
No log 9.6667 174 0.6375 0.7694 0.6375 0.7984
No log 9.7778 176 0.6416 0.7610 0.6416 0.8010
No log 9.8889 178 0.6448 0.7568 0.6448 0.8030
No log 10.0 180 0.6463 0.7500 0.6463 0.8039

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k4_task5_organization

Finetuned
(4023)
this model