ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k3_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6317
  • Qwk: 0.7478
  • Mse: 0.6317
  • Rmse: 0.7948

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1053 2 5.0300 0.0104 5.0300 2.2428
No log 0.2105 4 3.4484 0.0816 3.4484 1.8570
No log 0.3158 6 1.9537 0.1429 1.9537 1.3978
No log 0.4211 8 1.3386 0.2182 1.3386 1.1570
No log 0.5263 10 1.0294 0.2894 1.0294 1.0146
No log 0.6316 12 1.0124 0.2423 1.0124 1.0062
No log 0.7368 14 0.9383 0.3075 0.9383 0.9687
No log 0.8421 16 0.8835 0.3797 0.8835 0.9400
No log 0.9474 18 0.8324 0.5009 0.8324 0.9123
No log 1.0526 20 0.7614 0.5091 0.7614 0.8726
No log 1.1579 22 1.0157 0.5010 1.0157 1.0078
No log 1.2632 24 1.3209 0.4284 1.3209 1.1493
No log 1.3684 26 1.4267 0.3942 1.4267 1.1944
No log 1.4737 28 0.9740 0.6013 0.9740 0.9869
No log 1.5789 30 0.7838 0.7408 0.7838 0.8853
No log 1.6842 32 0.7764 0.7515 0.7764 0.8811
No log 1.7895 34 1.2608 0.4906 1.2608 1.1228
No log 1.8947 36 1.7894 0.3812 1.7894 1.3377
No log 2.0 38 1.3946 0.4635 1.3946 1.1809
No log 2.1053 40 1.0695 0.5954 1.0695 1.0342
No log 2.2105 42 0.6496 0.7285 0.6496 0.8060
No log 2.3158 44 0.5928 0.7639 0.5928 0.7699
No log 2.4211 46 0.6044 0.7396 0.6044 0.7774
No log 2.5263 48 0.8918 0.6386 0.8918 0.9443
No log 2.6316 50 1.0785 0.5714 1.0785 1.0385
No log 2.7368 52 0.8498 0.6660 0.8498 0.9219
No log 2.8421 54 0.6111 0.7610 0.6111 0.7818
No log 2.9474 56 0.6701 0.7572 0.6701 0.8186
No log 3.0526 58 0.8330 0.7221 0.8330 0.9127
No log 3.1579 60 0.7739 0.7365 0.7739 0.8797
No log 3.2632 62 0.5611 0.7645 0.5611 0.7491
No log 3.3684 64 0.5699 0.7400 0.5699 0.7549
No log 3.4737 66 0.7233 0.6551 0.7233 0.8505
No log 3.5789 68 0.7040 0.6746 0.7040 0.8391
No log 3.6842 70 0.6718 0.7047 0.6718 0.8196
No log 3.7895 72 0.5946 0.7654 0.5946 0.7711
No log 3.8947 74 0.6128 0.7723 0.6128 0.7828
No log 4.0 76 0.6559 0.7495 0.6559 0.8098
No log 4.1053 78 0.6436 0.7633 0.6436 0.8023
No log 4.2105 80 0.6010 0.7669 0.6010 0.7752
No log 4.3158 82 0.6147 0.7698 0.6147 0.7840
No log 4.4211 84 0.6468 0.7278 0.6468 0.8043
No log 4.5263 86 0.6148 0.7620 0.6148 0.7841
No log 4.6316 88 0.6005 0.7632 0.6005 0.7749
No log 4.7368 90 0.6770 0.7563 0.6770 0.8228
No log 4.8421 92 0.7320 0.7452 0.7320 0.8556
No log 4.9474 94 0.6612 0.7549 0.6612 0.8131
No log 5.0526 96 0.5885 0.7754 0.5885 0.7672
No log 5.1579 98 0.6548 0.7387 0.6548 0.8092
No log 5.2632 100 0.7356 0.7196 0.7356 0.8577
No log 5.3684 102 0.6760 0.7487 0.6760 0.8222
No log 5.4737 104 0.6136 0.7668 0.6136 0.7833
No log 5.5789 106 0.6918 0.7498 0.6918 0.8318
No log 5.6842 108 0.7124 0.7606 0.7124 0.8441
No log 5.7895 110 0.6328 0.7568 0.6328 0.7955
No log 5.8947 112 0.5844 0.7743 0.5844 0.7645
No log 6.0 114 0.5792 0.7563 0.5792 0.7611
No log 6.1053 116 0.5730 0.7562 0.5730 0.7570
No log 6.2105 118 0.5786 0.7831 0.5786 0.7607
No log 6.3158 120 0.5908 0.7732 0.5908 0.7686
No log 6.4211 122 0.6077 0.7802 0.6077 0.7795
No log 6.5263 124 0.6195 0.7592 0.6195 0.7871
No log 6.6316 126 0.6185 0.7528 0.6185 0.7864
No log 6.7368 128 0.6298 0.7573 0.6298 0.7936
No log 6.8421 130 0.6458 0.7571 0.6458 0.8036
No log 6.9474 132 0.6495 0.7646 0.6495 0.8059
No log 7.0526 134 0.6192 0.7560 0.6192 0.7869
No log 7.1579 136 0.5980 0.7692 0.5980 0.7733
No log 7.2632 138 0.6057 0.7682 0.6057 0.7783
No log 7.3684 140 0.6084 0.7601 0.6084 0.7800
No log 7.4737 142 0.6264 0.7730 0.6264 0.7914
No log 7.5789 144 0.6320 0.7734 0.6320 0.7950
No log 7.6842 146 0.6182 0.7617 0.6182 0.7862
No log 7.7895 148 0.6150 0.7833 0.6150 0.7842
No log 7.8947 150 0.6255 0.7766 0.6255 0.7909
No log 8.0 152 0.6424 0.7657 0.6424 0.8015
No log 8.1053 154 0.6638 0.7666 0.6638 0.8147
No log 8.2105 156 0.6642 0.7677 0.6642 0.8150
No log 8.3158 158 0.6563 0.7760 0.6563 0.8101
No log 8.4211 160 0.6566 0.7651 0.6566 0.8103
No log 8.5263 162 0.6603 0.7460 0.6603 0.8126
No log 8.6316 164 0.6609 0.7419 0.6609 0.8130
No log 8.7368 166 0.6578 0.7605 0.6578 0.8110
No log 8.8421 168 0.6457 0.7608 0.6457 0.8035
No log 8.9474 170 0.6329 0.7563 0.6329 0.7956
No log 9.0526 172 0.6259 0.7513 0.6259 0.7912
No log 9.1579 174 0.6230 0.7480 0.6230 0.7893
No log 9.2632 176 0.6248 0.7480 0.6248 0.7904
No log 9.3684 178 0.6322 0.7563 0.6322 0.7951
No log 9.4737 180 0.6375 0.7560 0.6375 0.7985
No log 9.5789 182 0.6394 0.7608 0.6394 0.7996
No log 9.6842 184 0.6379 0.7476 0.6379 0.7987
No log 9.7895 186 0.6355 0.7478 0.6355 0.7972
No log 9.8947 188 0.6329 0.7478 0.6329 0.7955
No log 10.0 190 0.6317 0.7478 0.6317 0.7948

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k3_task1_organization

Finetuned
(4023)
this model