ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k2_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8923
  • Qwk: 0.5146
  • Mse: 0.8923
  • Rmse: 0.9446

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 4.0144 0.0194 4.0144 2.0036
No log 0.3077 4 2.7717 0.0597 2.7717 1.6649
No log 0.4615 6 1.3383 0.0951 1.3383 1.1569
No log 0.6154 8 0.9628 0.0250 0.9628 0.9812
No log 0.7692 10 1.0457 -0.0802 1.0457 1.0226
No log 0.9231 12 0.7361 0.2069 0.7361 0.8579
No log 1.0769 14 0.7430 0.1469 0.7430 0.8620
No log 1.2308 16 0.7424 0.1261 0.7424 0.8616
No log 1.3846 18 0.7168 0.1777 0.7168 0.8466
No log 1.5385 20 0.6743 0.2108 0.6743 0.8211
No log 1.6923 22 0.6584 0.3128 0.6584 0.8114
No log 1.8462 24 0.6677 0.3270 0.6677 0.8171
No log 2.0 26 0.6739 0.3321 0.6739 0.8209
No log 2.1538 28 0.6089 0.3357 0.6089 0.7803
No log 2.3077 30 0.6385 0.2493 0.6385 0.7990
No log 2.4615 32 0.6599 0.2231 0.6599 0.8124
No log 2.6154 34 0.6449 0.2231 0.6449 0.8030
No log 2.7692 36 0.5909 0.3423 0.5909 0.7687
No log 2.9231 38 0.6063 0.3566 0.6063 0.7787
No log 3.0769 40 0.6448 0.3863 0.6448 0.8030
No log 3.2308 42 0.6881 0.3426 0.6881 0.8295
No log 3.3846 44 0.6065 0.3611 0.6065 0.7788
No log 3.5385 46 0.5506 0.4255 0.5506 0.7421
No log 3.6923 48 0.5805 0.4977 0.5805 0.7619
No log 3.8462 50 0.6709 0.5250 0.6709 0.8191
No log 4.0 52 0.7688 0.4631 0.7688 0.8768
No log 4.1538 54 0.7568 0.4758 0.7568 0.8699
No log 4.3077 56 0.6911 0.5194 0.6911 0.8313
No log 4.4615 58 0.6922 0.5178 0.6922 0.8320
No log 4.6154 60 0.7021 0.5224 0.7021 0.8379
No log 4.7692 62 0.7134 0.5219 0.7134 0.8446
No log 4.9231 64 0.7573 0.4912 0.7573 0.8702
No log 5.0769 66 0.7881 0.5042 0.7881 0.8878
No log 5.2308 68 0.8487 0.5370 0.8487 0.9213
No log 5.3846 70 0.8931 0.5274 0.8931 0.9450
No log 5.5385 72 0.9242 0.5164 0.9242 0.9613
No log 5.6923 74 0.9675 0.4608 0.9675 0.9836
No log 5.8462 76 0.9331 0.4626 0.9331 0.9660
No log 6.0 78 0.8574 0.4713 0.8574 0.9259
No log 6.1538 80 0.7944 0.5026 0.7944 0.8913
No log 6.3077 82 0.8113 0.5121 0.8113 0.9007
No log 6.4615 84 0.9064 0.4730 0.9064 0.9520
No log 6.6154 86 0.9760 0.4403 0.9760 0.9879
No log 6.7692 88 0.9304 0.4740 0.9304 0.9646
No log 6.9231 90 0.8666 0.4812 0.8666 0.9309
No log 7.0769 92 0.8457 0.5330 0.8457 0.9196
No log 7.2308 94 0.8679 0.4981 0.8679 0.9316
No log 7.3846 96 0.9112 0.4792 0.9112 0.9546
No log 7.5385 98 0.9357 0.4744 0.9357 0.9673
No log 7.6923 100 0.9652 0.4879 0.9652 0.9824
No log 7.8462 102 0.9416 0.4652 0.9416 0.9703
No log 8.0 104 0.9040 0.4960 0.9040 0.9508
No log 8.1538 106 0.8942 0.5062 0.8942 0.9456
No log 8.3077 108 0.8924 0.4859 0.8924 0.9447
No log 8.4615 110 0.8840 0.5021 0.8840 0.9402
No log 8.6154 112 0.8787 0.4954 0.8787 0.9374
No log 8.7692 114 0.8806 0.5014 0.8806 0.9384
No log 8.9231 116 0.9005 0.5086 0.9005 0.9490
No log 9.0769 118 0.9086 0.4884 0.9086 0.9532
No log 9.2308 120 0.9190 0.5016 0.9190 0.9586
No log 9.3846 122 0.9130 0.4977 0.9130 0.9555
No log 9.5385 124 0.9051 0.5213 0.9051 0.9514
No log 9.6923 126 0.8954 0.5086 0.8954 0.9463
No log 9.8462 128 0.8934 0.5086 0.8934 0.9452
No log 10.0 130 0.8923 0.5146 0.8923 0.9446

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k2_task2_organization

Finetuned
(4023)
this model