ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k3_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0355
  • Qwk: 0.4354
  • Mse: 1.0355
  • Rmse: 1.0176

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1053 2 4.1530 0.0077 4.1530 2.0379
No log 0.2105 4 2.3575 0.0230 2.3575 1.5354
No log 0.3158 6 1.4268 0.0324 1.4268 1.1945
No log 0.4211 8 0.9886 -0.0238 0.9886 0.9943
No log 0.5263 10 0.7806 0.1523 0.7806 0.8835
No log 0.6316 12 0.7993 0.1819 0.7993 0.8940
No log 0.7368 14 0.7799 0.0831 0.7799 0.8831
No log 0.8421 16 0.9977 -0.0455 0.9977 0.9988
No log 0.9474 18 1.0012 0.0080 1.0012 1.0006
No log 1.0526 20 0.7831 0.0574 0.7831 0.8849
No log 1.1579 22 0.7498 0.2326 0.7498 0.8659
No log 1.2632 24 0.8287 0.3176 0.8287 0.9103
No log 1.3684 26 0.7626 0.3480 0.7626 0.8733
No log 1.4737 28 0.6360 0.2729 0.6360 0.7975
No log 1.5789 30 0.5664 0.3033 0.5664 0.7526
No log 1.6842 32 0.5598 0.3423 0.5598 0.7482
No log 1.7895 34 0.5706 0.4119 0.5706 0.7554
No log 1.8947 36 0.5896 0.4081 0.5896 0.7679
No log 2.0 38 0.6676 0.4097 0.6676 0.8171
No log 2.1053 40 0.6277 0.4916 0.6277 0.7923
No log 2.2105 42 0.7472 0.4633 0.7472 0.8644
No log 2.3158 44 0.8005 0.4692 0.8005 0.8947
No log 2.4211 46 0.6579 0.4722 0.6579 0.8111
No log 2.5263 48 0.7521 0.4308 0.7521 0.8672
No log 2.6316 50 1.1557 0.3299 1.1557 1.0750
No log 2.7368 52 1.3369 0.2579 1.3369 1.1562
No log 2.8421 54 1.0450 0.3482 1.0450 1.0222
No log 2.9474 56 0.6684 0.4835 0.6684 0.8176
No log 3.0526 58 0.6043 0.4427 0.6043 0.7774
No log 3.1579 60 0.6240 0.4615 0.6240 0.7900
No log 3.2632 62 0.6865 0.4792 0.6865 0.8286
No log 3.3684 64 0.8024 0.4328 0.8024 0.8958
No log 3.4737 66 0.8555 0.4564 0.8555 0.9249
No log 3.5789 68 1.0808 0.3971 1.0808 1.0396
No log 3.6842 70 1.1555 0.4008 1.1555 1.0749
No log 3.7895 72 0.9095 0.4253 0.9095 0.9537
No log 3.8947 74 0.7523 0.5256 0.7523 0.8674
No log 4.0 76 0.8509 0.4961 0.8509 0.9224
No log 4.1053 78 0.8724 0.4902 0.8724 0.9340
No log 4.2105 80 0.8180 0.5666 0.8180 0.9045
No log 4.3158 82 0.8785 0.5191 0.8785 0.9373
No log 4.4211 84 0.9703 0.4998 0.9703 0.9850
No log 4.5263 86 1.0372 0.4645 1.0372 1.0185
No log 4.6316 88 1.1185 0.4326 1.1185 1.0576
No log 4.7368 90 1.0877 0.4383 1.0877 1.0429
No log 4.8421 92 0.9995 0.4657 0.9995 0.9998
No log 4.9474 94 0.9317 0.4673 0.9317 0.9653
No log 5.0526 96 0.8710 0.4846 0.8710 0.9333
No log 5.1579 98 0.8622 0.4813 0.8622 0.9286
No log 5.2632 100 0.8841 0.4785 0.8841 0.9402
No log 5.3684 102 0.9933 0.4472 0.9933 0.9966
No log 5.4737 104 1.0149 0.4466 1.0149 1.0074
No log 5.5789 106 0.9408 0.4763 0.9408 0.9699
No log 5.6842 108 0.9390 0.4936 0.9390 0.9690
No log 5.7895 110 0.8668 0.5081 0.8668 0.9310
No log 5.8947 112 0.8482 0.5018 0.8482 0.9210
No log 6.0 114 0.7986 0.5092 0.7986 0.8936
No log 6.1053 116 0.8158 0.5355 0.8158 0.9032
No log 6.2105 118 0.8544 0.5438 0.8544 0.9243
No log 6.3158 120 0.8750 0.5477 0.8750 0.9354
No log 6.4211 122 0.9111 0.5419 0.9111 0.9545
No log 6.5263 124 0.9815 0.5192 0.9815 0.9907
No log 6.6316 126 1.0505 0.4928 1.0505 1.0250
No log 6.7368 128 1.0296 0.5198 1.0296 1.0147
No log 6.8421 130 0.9474 0.5316 0.9474 0.9733
No log 6.9474 132 0.9063 0.5380 0.9063 0.9520
No log 7.0526 134 0.9000 0.5301 0.9000 0.9487
No log 7.1579 136 0.8935 0.5282 0.8935 0.9453
No log 7.2632 138 0.9428 0.4844 0.9428 0.9710
No log 7.3684 140 0.9670 0.4808 0.9670 0.9834
No log 7.4737 142 0.9732 0.4808 0.9732 0.9865
No log 7.5789 144 1.0218 0.4446 1.0218 1.0108
No log 7.6842 146 0.9971 0.4724 0.9971 0.9985
No log 7.7895 148 0.9684 0.4724 0.9684 0.9841
No log 7.8947 150 0.9062 0.5101 0.9062 0.9519
No log 8.0 152 0.8692 0.5018 0.8692 0.9323
No log 8.1053 154 0.8393 0.5106 0.8393 0.9161
No log 8.2105 156 0.8104 0.5114 0.8104 0.9002
No log 8.3158 158 0.8067 0.5323 0.8067 0.8982
No log 8.4211 160 0.8185 0.5338 0.8185 0.9047
No log 8.5263 162 0.8420 0.5312 0.8420 0.9176
No log 8.6316 164 0.8733 0.5436 0.8733 0.9345
No log 8.7368 166 0.9081 0.5136 0.9081 0.9530
No log 8.8421 168 0.9269 0.5169 0.9269 0.9627
No log 8.9474 170 0.9520 0.4821 0.9520 0.9757
No log 9.0526 172 0.9966 0.4851 0.9966 0.9983
No log 9.1579 174 1.0600 0.4439 1.0600 1.0295
No log 9.2632 176 1.1138 0.4070 1.1138 1.0553
No log 9.3684 178 1.1350 0.4015 1.1350 1.0654
No log 9.4737 180 1.1290 0.4015 1.1290 1.0625
No log 9.5789 182 1.1021 0.4109 1.1021 1.0498
No log 9.6842 184 1.0708 0.4225 1.0708 1.0348
No log 9.7895 186 1.0546 0.4495 1.0546 1.0269
No log 9.8947 188 1.0414 0.4354 1.0414 1.0205
No log 10.0 190 1.0355 0.4354 1.0355 1.0176

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k3_task2_organization

Finetuned
(4019)
this model