ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k2_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7546
  • Qwk: 0.7354
  • Mse: 0.7546
  • Rmse: 0.8687

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1333 2 5.1824 -0.0323 5.1824 2.2765
No log 0.2667 4 3.6030 0.0707 3.6030 1.8982
No log 0.4 6 2.0550 0.1300 2.0550 1.4335
No log 0.5333 8 1.8337 0.0019 1.8337 1.3541
No log 0.6667 10 1.6659 0.0361 1.6659 1.2907
No log 0.8 12 1.3265 0.1447 1.3265 1.1517
No log 0.9333 14 1.2051 0.1688 1.2051 1.0977
No log 1.0667 16 1.2359 0.1767 1.2359 1.1117
No log 1.2 18 1.0991 0.3203 1.0991 1.0484
No log 1.3333 20 0.9935 0.4469 0.9935 0.9967
No log 1.4667 22 1.0461 0.4449 1.0461 1.0228
No log 1.6 24 1.0640 0.4361 1.0640 1.0315
No log 1.7333 26 0.8554 0.5456 0.8554 0.9249
No log 1.8667 28 0.8140 0.5341 0.8140 0.9022
No log 2.0 30 0.7995 0.5140 0.7995 0.8942
No log 2.1333 32 0.8216 0.4881 0.8216 0.9064
No log 2.2667 34 0.8021 0.5192 0.8021 0.8956
No log 2.4 36 0.8038 0.5480 0.8038 0.8966
No log 2.5333 38 0.8426 0.5590 0.8426 0.9179
No log 2.6667 40 0.9434 0.5055 0.9434 0.9713
No log 2.8 42 0.8122 0.6420 0.8122 0.9012
No log 2.9333 44 0.7790 0.6365 0.7790 0.8826
No log 3.0667 46 0.7981 0.6084 0.7981 0.8934
No log 3.2 48 0.7567 0.6459 0.7567 0.8699
No log 3.3333 50 0.8096 0.6704 0.8096 0.8998
No log 3.4667 52 0.8975 0.6454 0.8975 0.9474
No log 3.6 54 0.8738 0.6486 0.8738 0.9348
No log 3.7333 56 0.7492 0.6870 0.7492 0.8656
No log 3.8667 58 0.7177 0.6609 0.7177 0.8472
No log 4.0 60 0.7220 0.6773 0.7220 0.8497
No log 4.1333 62 0.7891 0.6835 0.7891 0.8883
No log 4.2667 64 0.7116 0.6963 0.7116 0.8435
No log 4.4 66 0.7464 0.6928 0.7464 0.8640
No log 4.5333 68 0.7899 0.6833 0.7899 0.8887
No log 4.6667 70 0.6895 0.7051 0.6895 0.8304
No log 4.8 72 0.7664 0.7202 0.7664 0.8754
No log 4.9333 74 1.1343 0.5577 1.1343 1.0650
No log 5.0667 76 1.2956 0.4826 1.2956 1.1382
No log 5.2 78 1.1725 0.5325 1.1725 1.0828
No log 5.3333 80 0.8672 0.7006 0.8672 0.9312
No log 5.4667 82 0.6691 0.7084 0.6691 0.8180
No log 5.6 84 0.6726 0.6943 0.6726 0.8201
No log 5.7333 86 0.6693 0.6942 0.6693 0.8181
No log 5.8667 88 0.6796 0.7100 0.6796 0.8244
No log 6.0 90 0.7565 0.7273 0.7565 0.8698
No log 6.1333 92 0.9236 0.6631 0.9236 0.9611
No log 6.2667 94 0.9062 0.6631 0.9062 0.9519
No log 6.4 96 0.7624 0.7328 0.7624 0.8732
No log 6.5333 98 0.6519 0.7273 0.6519 0.8074
No log 6.6667 100 0.6738 0.6890 0.6738 0.8209
No log 6.8 102 0.6940 0.6872 0.6940 0.8331
No log 6.9333 104 0.6509 0.6998 0.6509 0.8068
No log 7.0667 106 0.6345 0.7175 0.6345 0.7965
No log 7.2 108 0.6947 0.7391 0.6947 0.8335
No log 7.3333 110 0.8361 0.6968 0.8361 0.9144
No log 7.4667 112 0.8875 0.6820 0.8875 0.9420
No log 7.6 114 0.9409 0.6397 0.9409 0.9700
No log 7.7333 116 0.8663 0.6790 0.8663 0.9308
No log 7.8667 118 0.7662 0.7349 0.7662 0.8753
No log 8.0 120 0.6960 0.7485 0.6960 0.8343
No log 8.1333 122 0.6706 0.7539 0.6706 0.8189
No log 8.2667 124 0.6755 0.7576 0.6755 0.8219
No log 8.4 126 0.7013 0.7488 0.7013 0.8374
No log 8.5333 128 0.7386 0.7298 0.7386 0.8594
No log 8.6667 130 0.7691 0.7328 0.7691 0.8770
No log 8.8 132 0.8195 0.7084 0.8195 0.9053
No log 8.9333 134 0.8380 0.7047 0.8380 0.9154
No log 9.0667 136 0.8231 0.7176 0.8231 0.9073
No log 9.2 138 0.7941 0.7363 0.7941 0.8911
No log 9.3333 140 0.7670 0.7389 0.7670 0.8758
No log 9.4667 142 0.7538 0.7354 0.7538 0.8682
No log 9.6 144 0.7484 0.7420 0.7484 0.8651
No log 9.7333 146 0.7527 0.7319 0.7527 0.8676
No log 9.8667 148 0.7536 0.7319 0.7536 0.8681
No log 10.0 150 0.7546 0.7354 0.7546 0.8687

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k2_task1_organization

Finetuned
(4023)
this model