ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k8_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6051
  • Qwk: 0.7376
  • Mse: 0.6051
  • Rmse: 0.7779

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0476 2 5.2938 -0.0030 5.2938 2.3008
No log 0.0952 4 3.3036 0.0699 3.3036 1.8176
No log 0.1429 6 2.3015 -0.0002 2.3015 1.5171
No log 0.1905 8 1.8429 -0.0149 1.8429 1.3575
No log 0.2381 10 1.2653 0.2564 1.2653 1.1248
No log 0.2857 12 1.1630 0.2958 1.1630 1.0784
No log 0.3333 14 1.3791 0.1865 1.3791 1.1743
No log 0.3810 16 1.2627 0.2408 1.2627 1.1237
No log 0.4286 18 1.5268 0.0886 1.5268 1.2356
No log 0.4762 20 2.0545 0.1298 2.0545 1.4334
No log 0.5238 22 1.6676 0.1297 1.6676 1.2914
No log 0.5714 24 1.1985 0.2177 1.1985 1.0947
No log 0.6190 26 1.1435 0.3041 1.1435 1.0694
No log 0.6667 28 0.9467 0.4865 0.9467 0.9730
No log 0.7143 30 0.9885 0.5423 0.9885 0.9942
No log 0.7619 32 1.5731 0.3427 1.5731 1.2543
No log 0.8095 34 1.8906 0.3134 1.8906 1.3750
No log 0.8571 36 1.6641 0.3574 1.6641 1.2900
No log 0.9048 38 0.9390 0.5854 0.9390 0.9690
No log 0.9524 40 0.6081 0.6798 0.6081 0.7798
No log 1.0 42 0.7510 0.5309 0.7510 0.8666
No log 1.0476 44 0.7341 0.4665 0.7341 0.8568
No log 1.0952 46 0.6957 0.5714 0.6957 0.8341
No log 1.1429 48 0.9038 0.4701 0.9038 0.9507
No log 1.1905 50 1.3039 0.3862 1.3039 1.1419
No log 1.2381 52 1.4056 0.3316 1.4056 1.1856
No log 1.2857 54 1.2141 0.4495 1.2141 1.1019
No log 1.3333 56 0.9058 0.5512 0.9058 0.9517
No log 1.3810 58 0.7807 0.5948 0.7807 0.8836
No log 1.4286 60 0.7169 0.6391 0.7169 0.8467
No log 1.4762 62 0.7853 0.6409 0.7853 0.8862
No log 1.5238 64 0.7242 0.6932 0.7242 0.8510
No log 1.5714 66 1.0116 0.5860 1.0116 1.0058
No log 1.6190 68 1.5052 0.3973 1.5052 1.2269
No log 1.6667 70 1.9954 0.2912 1.9954 1.4126
No log 1.7143 72 1.8634 0.3213 1.8634 1.3651
No log 1.7619 74 1.2083 0.4680 1.2083 1.0992
No log 1.8095 76 0.6817 0.6835 0.6817 0.8256
No log 1.8571 78 0.5686 0.6925 0.5686 0.7541
No log 1.9048 80 0.5743 0.7152 0.5743 0.7578
No log 1.9524 82 0.6900 0.6546 0.6900 0.8306
No log 2.0 84 0.9642 0.5579 0.9642 0.9819
No log 2.0476 86 1.2134 0.4634 1.2134 1.1016
No log 2.0952 88 1.0828 0.5110 1.0828 1.0406
No log 2.1429 90 0.7754 0.6412 0.7754 0.8806
No log 2.1905 92 0.5839 0.7102 0.5839 0.7641
No log 2.2381 94 0.5702 0.7484 0.5702 0.7551
No log 2.2857 96 0.5819 0.7429 0.5819 0.7628
No log 2.3333 98 0.6034 0.7113 0.6034 0.7768
No log 2.3810 100 0.6153 0.7147 0.6153 0.7844
No log 2.4286 102 0.6042 0.7239 0.6042 0.7773
No log 2.4762 104 0.6312 0.7078 0.6312 0.7945
No log 2.5238 106 0.7659 0.6621 0.7659 0.8751
No log 2.5714 108 0.7944 0.6681 0.7944 0.8913
No log 2.6190 110 0.7613 0.6767 0.7613 0.8725
No log 2.6667 112 0.7539 0.6767 0.7539 0.8683
No log 2.7143 114 0.6441 0.6911 0.6441 0.8026
No log 2.7619 116 0.6313 0.6950 0.6313 0.7945
No log 2.8095 118 0.6553 0.6810 0.6553 0.8095
No log 2.8571 120 0.6361 0.6791 0.6361 0.7976
No log 2.9048 122 0.5770 0.7155 0.5770 0.7596
No log 2.9524 124 0.5657 0.7709 0.5657 0.7521
No log 3.0 126 0.5815 0.7573 0.5815 0.7625
No log 3.0476 128 0.5968 0.7470 0.5968 0.7725
No log 3.0952 130 0.6117 0.7355 0.6117 0.7821
No log 3.1429 132 0.6783 0.7371 0.6783 0.8236
No log 3.1905 134 0.7412 0.7249 0.7412 0.8609
No log 3.2381 136 0.7337 0.7313 0.7337 0.8566
No log 3.2857 138 0.6469 0.7205 0.6469 0.8043
No log 3.3333 140 0.6356 0.7281 0.6356 0.7972
No log 3.3810 142 0.6452 0.7285 0.6452 0.8033
No log 3.4286 144 0.6094 0.7269 0.6094 0.7806
No log 3.4762 146 0.6838 0.7150 0.6838 0.8270
No log 3.5238 148 0.6602 0.7245 0.6602 0.8125
No log 3.5714 150 0.6113 0.7350 0.6113 0.7819
No log 3.6190 152 0.8264 0.6478 0.8264 0.9091
No log 3.6667 154 1.0101 0.6071 1.0101 1.0050
No log 3.7143 156 0.9152 0.6274 0.9152 0.9567
No log 3.7619 158 0.6949 0.6889 0.6949 0.8336
No log 3.8095 160 0.5845 0.7583 0.5845 0.7645
No log 3.8571 162 0.5856 0.7347 0.5856 0.7652
No log 3.9048 164 0.5685 0.7588 0.5685 0.7540
No log 3.9524 166 0.6135 0.7275 0.6135 0.7833
No log 4.0 168 0.6688 0.6999 0.6688 0.8178
No log 4.0476 170 0.6608 0.7077 0.6608 0.8129
No log 4.0952 172 0.5867 0.7260 0.5867 0.7660
No log 4.1429 174 0.5784 0.7371 0.5784 0.7605
No log 4.1905 176 0.5842 0.7350 0.5842 0.7643
No log 4.2381 178 0.6136 0.7280 0.6136 0.7833
No log 4.2857 180 0.6436 0.7159 0.6436 0.8023
No log 4.3333 182 0.6748 0.6991 0.6748 0.8215
No log 4.3810 184 0.6589 0.6793 0.6589 0.8117
No log 4.4286 186 0.6124 0.7306 0.6124 0.7825
No log 4.4762 188 0.5719 0.7421 0.5719 0.7562
No log 4.5238 190 0.5586 0.7395 0.5586 0.7474
No log 4.5714 192 0.5542 0.7384 0.5542 0.7444
No log 4.6190 194 0.5579 0.7512 0.5579 0.7469
No log 4.6667 196 0.5618 0.7443 0.5618 0.7495
No log 4.7143 198 0.5692 0.7419 0.5692 0.7545
No log 4.7619 200 0.5845 0.7331 0.5845 0.7645
No log 4.8095 202 0.5647 0.7284 0.5647 0.7515
No log 4.8571 204 0.5678 0.7447 0.5678 0.7535
No log 4.9048 206 0.5675 0.7413 0.5675 0.7533
No log 4.9524 208 0.5818 0.7477 0.5818 0.7627
No log 5.0 210 0.5760 0.7526 0.5760 0.7589
No log 5.0476 212 0.6013 0.7097 0.6013 0.7755
No log 5.0952 214 0.6402 0.7044 0.6402 0.8001
No log 5.1429 216 0.6140 0.6979 0.6140 0.7836
No log 5.1905 218 0.5934 0.7373 0.5934 0.7703
No log 5.2381 220 0.5910 0.7293 0.5910 0.7688
No log 5.2857 222 0.5887 0.7373 0.5887 0.7673
No log 5.3333 224 0.5844 0.7363 0.5844 0.7645
No log 5.3810 226 0.5892 0.7142 0.5892 0.7676
No log 5.4286 228 0.5972 0.7146 0.5972 0.7728
No log 5.4762 230 0.5968 0.7338 0.5968 0.7726
No log 5.5238 232 0.5658 0.7243 0.5658 0.7522
No log 5.5714 234 0.5659 0.7523 0.5659 0.7523
No log 5.6190 236 0.5843 0.7434 0.5843 0.7644
No log 5.6667 238 0.5888 0.7581 0.5888 0.7673
No log 5.7143 240 0.6006 0.7193 0.6006 0.7750
No log 5.7619 242 0.6320 0.7036 0.6320 0.7950
No log 5.8095 244 0.6652 0.6865 0.6652 0.8156
No log 5.8571 246 0.6344 0.6966 0.6344 0.7965
No log 5.9048 248 0.5951 0.7046 0.5951 0.7714
No log 5.9524 250 0.5870 0.7122 0.5870 0.7661
No log 6.0 252 0.5829 0.7377 0.5829 0.7635
No log 6.0476 254 0.5957 0.7314 0.5957 0.7718
No log 6.0952 256 0.5858 0.7248 0.5858 0.7654
No log 6.1429 258 0.5883 0.7135 0.5883 0.7670
No log 6.1905 260 0.6228 0.6853 0.6228 0.7892
No log 6.2381 262 0.6383 0.6740 0.6383 0.7990
No log 6.2857 264 0.6106 0.6854 0.6106 0.7814
No log 6.3333 266 0.5992 0.7277 0.5992 0.7741
No log 6.3810 268 0.6170 0.7449 0.6170 0.7855
No log 6.4286 270 0.6285 0.7329 0.6285 0.7928
No log 6.4762 272 0.6234 0.7329 0.6234 0.7896
No log 6.5238 274 0.6062 0.7566 0.6062 0.7786
No log 6.5714 276 0.6120 0.7386 0.6120 0.7823
No log 6.6190 278 0.6090 0.7455 0.6090 0.7804
No log 6.6667 280 0.5963 0.7287 0.5963 0.7722
No log 6.7143 282 0.5942 0.6976 0.5942 0.7708
No log 6.7619 284 0.5978 0.6964 0.5978 0.7732
No log 6.8095 286 0.5915 0.6976 0.5915 0.7691
No log 6.8571 288 0.5997 0.7144 0.5997 0.7744
No log 6.9048 290 0.6131 0.7352 0.6131 0.7830
No log 6.9524 292 0.6136 0.7364 0.6136 0.7833
No log 7.0 294 0.6006 0.7248 0.6006 0.7750
No log 7.0476 296 0.5899 0.7033 0.5899 0.7681
No log 7.0952 298 0.5885 0.7066 0.5885 0.7671
No log 7.1429 300 0.6022 0.6921 0.6022 0.7760
No log 7.1905 302 0.6040 0.6921 0.6040 0.7772
No log 7.2381 304 0.5910 0.6888 0.5910 0.7688
No log 7.2857 306 0.5798 0.7047 0.5798 0.7615
No log 7.3333 308 0.5782 0.7072 0.5782 0.7604
No log 7.3810 310 0.5900 0.7325 0.5900 0.7681
No log 7.4286 312 0.5894 0.7325 0.5894 0.7677
No log 7.4762 314 0.5832 0.7380 0.5832 0.7637
No log 7.5238 316 0.5777 0.7223 0.5777 0.7601
No log 7.5714 318 0.5843 0.7163 0.5843 0.7644
No log 7.6190 320 0.5838 0.7060 0.5838 0.7641
No log 7.6667 322 0.5799 0.7127 0.5799 0.7615
No log 7.7143 324 0.5767 0.7228 0.5767 0.7594
No log 7.7619 326 0.5766 0.7189 0.5766 0.7593
No log 7.8095 328 0.5794 0.6978 0.5794 0.7612
No log 7.8571 330 0.5829 0.6919 0.5829 0.7635
No log 7.9048 332 0.5815 0.7190 0.5815 0.7626
No log 7.9524 334 0.5840 0.7435 0.5840 0.7642
No log 8.0 336 0.5886 0.7366 0.5886 0.7672
No log 8.0476 338 0.5919 0.7366 0.5919 0.7693
No log 8.0952 340 0.5993 0.7291 0.5993 0.7742
No log 8.1429 342 0.6105 0.7354 0.6105 0.7813
No log 8.1905 344 0.6183 0.7370 0.6183 0.7863
No log 8.2381 346 0.6206 0.7302 0.6206 0.7878
No log 8.2857 348 0.6194 0.7370 0.6194 0.7870
No log 8.3333 350 0.6155 0.7354 0.6155 0.7845
No log 8.3810 352 0.6221 0.7416 0.6221 0.7887
No log 8.4286 354 0.6240 0.7416 0.6240 0.7900
No log 8.4762 356 0.6197 0.7461 0.6197 0.7872
No log 8.5238 358 0.6132 0.7366 0.6132 0.7830
No log 8.5714 360 0.6126 0.7366 0.6126 0.7827
No log 8.6190 362 0.6124 0.7366 0.6124 0.7826
No log 8.6667 364 0.6103 0.7366 0.6103 0.7812
No log 8.7143 366 0.6098 0.7236 0.6098 0.7809
No log 8.7619 368 0.6092 0.7236 0.6092 0.7805
No log 8.8095 370 0.6077 0.7236 0.6077 0.7795
No log 8.8571 372 0.6073 0.7236 0.6073 0.7793
No log 8.9048 374 0.6085 0.7221 0.6085 0.7801
No log 8.9524 376 0.6096 0.7285 0.6096 0.7808
No log 9.0 378 0.6092 0.7285 0.6092 0.7805
No log 9.0476 380 0.6115 0.7285 0.6115 0.7820
No log 9.0952 382 0.6124 0.7285 0.6124 0.7826
No log 9.1429 384 0.6174 0.7302 0.6174 0.7857
No log 9.1905 386 0.6188 0.7302 0.6188 0.7866
No log 9.2381 388 0.6178 0.7302 0.6178 0.7860
No log 9.2857 390 0.6138 0.7348 0.6138 0.7835
No log 9.3333 392 0.6073 0.7183 0.6073 0.7793
No log 9.3810 394 0.6022 0.7251 0.6022 0.7760
No log 9.4286 396 0.5997 0.7366 0.5997 0.7744
No log 9.4762 398 0.5979 0.7366 0.5979 0.7732
No log 9.5238 400 0.5976 0.7249 0.5976 0.7730
No log 9.5714 402 0.5974 0.7249 0.5974 0.7729
No log 9.6190 404 0.5978 0.7249 0.5978 0.7732
No log 9.6667 406 0.5993 0.7249 0.5993 0.7741
No log 9.7143 408 0.6009 0.7349 0.6009 0.7752
No log 9.7619 410 0.6021 0.7366 0.6021 0.7759
No log 9.8095 412 0.6031 0.7428 0.6031 0.7766
No log 9.8571 414 0.6042 0.7428 0.6042 0.7773
No log 9.9048 416 0.6051 0.7314 0.6051 0.7779
No log 9.9524 418 0.6052 0.7376 0.6052 0.7779
No log 10.0 420 0.6051 0.7376 0.6051 0.7779

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run2_AugV5_k8_task1_organization

Finetuned
(4023)
this model