ArabicNewSplits6_FineTuningAraBERT_run1_AugV5_k6_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5932
  • Qwk: 0.7465
  • Mse: 0.5932
  • Rmse: 0.7702

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0606 2 5.0059 -0.0196 5.0059 2.2374
No log 0.1212 4 2.9400 0.0955 2.9400 1.7146
No log 0.1818 6 2.0699 0.0474 2.0699 1.4387
No log 0.2424 8 1.4859 0.0691 1.4859 1.2190
No log 0.3030 10 1.2699 0.1910 1.2699 1.1269
No log 0.3636 12 1.2085 0.2358 1.2085 1.0993
No log 0.4242 14 1.2801 0.1889 1.2801 1.1314
No log 0.4848 16 1.2942 0.1865 1.2942 1.1376
No log 0.5455 18 1.2509 0.2828 1.2509 1.1184
No log 0.6061 20 1.0644 0.1879 1.0644 1.0317
No log 0.6667 22 1.0884 0.1899 1.0884 1.0432
No log 0.7273 24 0.9375 0.3231 0.9375 0.9683
No log 0.7879 26 1.0021 0.4775 1.0021 1.0011
No log 0.8485 28 2.0234 0.2640 2.0234 1.4225
No log 0.9091 30 2.6116 0.1530 2.6116 1.6160
No log 0.9697 32 2.2942 0.2373 2.2942 1.5147
No log 1.0303 34 1.6186 0.2419 1.6186 1.2723
No log 1.0909 36 1.1085 0.3678 1.1085 1.0529
No log 1.1515 38 1.0277 0.3926 1.0277 1.0138
No log 1.2121 40 0.9474 0.4141 0.9474 0.9734
No log 1.2727 42 0.8716 0.4664 0.8716 0.9336
No log 1.3333 44 0.8668 0.4735 0.8668 0.9310
No log 1.3939 46 0.9222 0.4747 0.9222 0.9603
No log 1.4545 48 0.8581 0.5036 0.8581 0.9263
No log 1.5152 50 0.8856 0.5054 0.8856 0.9411
No log 1.5758 52 0.8984 0.4420 0.8984 0.9478
No log 1.6364 54 0.8144 0.5537 0.8144 0.9024
No log 1.6970 56 0.9753 0.4882 0.9753 0.9875
No log 1.7576 58 0.8759 0.5659 0.8759 0.9359
No log 1.8182 60 0.6779 0.6973 0.6779 0.8233
No log 1.8788 62 0.7007 0.6809 0.7007 0.8371
No log 1.9394 64 0.9889 0.5947 0.9889 0.9944
No log 2.0 66 1.3893 0.4556 1.3893 1.1787
No log 2.0606 68 1.2972 0.4551 1.2972 1.1390
No log 2.1212 70 0.8943 0.5914 0.8943 0.9457
No log 2.1818 72 0.6224 0.7028 0.6224 0.7889
No log 2.2424 74 0.5780 0.7626 0.5780 0.7603
No log 2.3030 76 0.6029 0.7144 0.6029 0.7764
No log 2.3636 78 0.7294 0.6743 0.7294 0.8541
No log 2.4242 80 0.6626 0.6848 0.6626 0.8140
No log 2.4848 82 0.5826 0.7460 0.5826 0.7633
No log 2.5455 84 0.5542 0.7751 0.5542 0.7444
No log 2.6061 86 0.5575 0.7709 0.5575 0.7467
No log 2.6667 88 0.5570 0.7857 0.5570 0.7464
No log 2.7273 90 0.5936 0.7504 0.5936 0.7704
No log 2.7879 92 0.7929 0.6667 0.7929 0.8904
No log 2.8485 94 0.9683 0.5769 0.9683 0.9840
No log 2.9091 96 0.9467 0.5641 0.9467 0.9730
No log 2.9697 98 0.7698 0.6614 0.7698 0.8774
No log 3.0303 100 0.5949 0.7143 0.5949 0.7713
No log 3.0909 102 0.5646 0.7106 0.5646 0.7514
No log 3.1515 104 0.5651 0.7240 0.5651 0.7517
No log 3.2121 106 0.5574 0.7264 0.5574 0.7466
No log 3.2727 108 0.5841 0.7227 0.5841 0.7643
No log 3.3333 110 0.7378 0.6833 0.7378 0.8589
No log 3.3939 112 0.9188 0.5951 0.9188 0.9585
No log 3.4545 114 0.9037 0.6037 0.9037 0.9506
No log 3.5152 116 0.7420 0.6795 0.7420 0.8614
No log 3.5758 118 0.5831 0.7393 0.5831 0.7636
No log 3.6364 120 0.5919 0.7384 0.5919 0.7693
No log 3.6970 122 0.6234 0.7502 0.6234 0.7896
No log 3.7576 124 0.6106 0.7412 0.6106 0.7814
No log 3.8182 126 0.6180 0.7470 0.6180 0.7861
No log 3.8788 128 0.6175 0.7571 0.6175 0.7858
No log 3.9394 130 0.6525 0.7464 0.6525 0.8078
No log 4.0 132 0.7098 0.7110 0.7098 0.8425
No log 4.0606 134 0.6610 0.7502 0.6610 0.8130
No log 4.1212 136 0.5986 0.7373 0.5986 0.7737
No log 4.1818 138 0.5895 0.7548 0.5895 0.7678
No log 4.2424 140 0.5854 0.7644 0.5853 0.7651
No log 4.3030 142 0.5923 0.7448 0.5923 0.7696
No log 4.3636 144 0.5778 0.7708 0.5778 0.7601
No log 4.4242 146 0.5683 0.7629 0.5683 0.7539
No log 4.4848 148 0.6109 0.7459 0.6109 0.7816
No log 4.5455 150 0.6066 0.7564 0.6066 0.7788
No log 4.6061 152 0.5919 0.7684 0.5919 0.7693
No log 4.6667 154 0.6176 0.7641 0.6176 0.7859
No log 4.7273 156 0.6607 0.7439 0.6607 0.8128
No log 4.7879 158 0.6817 0.7330 0.6817 0.8256
No log 4.8485 160 0.6608 0.7488 0.6608 0.8129
No log 4.9091 162 0.6506 0.7460 0.6506 0.8066
No log 4.9697 164 0.6780 0.7426 0.6780 0.8234
No log 5.0303 166 0.7511 0.7059 0.7511 0.8667
No log 5.0909 168 0.7693 0.7112 0.7693 0.8771
No log 5.1515 170 0.7160 0.7316 0.7160 0.8461
No log 5.2121 172 0.6618 0.7443 0.6618 0.8135
No log 5.2727 174 0.6308 0.7540 0.6308 0.7942
No log 5.3333 176 0.6183 0.7552 0.6183 0.7863
No log 5.3939 178 0.6070 0.7662 0.6070 0.7791
No log 5.4545 180 0.6213 0.7437 0.6213 0.7882
No log 5.5152 182 0.6249 0.7437 0.6249 0.7905
No log 5.5758 184 0.6132 0.7583 0.6132 0.7831
No log 5.6364 186 0.6246 0.7583 0.6246 0.7903
No log 5.6970 188 0.6036 0.7619 0.6036 0.7769
No log 5.7576 190 0.5821 0.7604 0.5821 0.7629
No log 5.8182 192 0.5722 0.7539 0.5722 0.7565
No log 5.8788 194 0.5714 0.7504 0.5714 0.7559
No log 5.9394 196 0.5824 0.7649 0.5824 0.7631
No log 6.0 198 0.6068 0.7695 0.6068 0.7790
No log 6.0606 200 0.6087 0.7609 0.6087 0.7802
No log 6.1212 202 0.5968 0.75 0.5968 0.7725
No log 6.1818 204 0.5760 0.7481 0.5760 0.7589
No log 6.2424 206 0.5764 0.7628 0.5764 0.7592
No log 6.3030 208 0.5856 0.7477 0.5856 0.7652
No log 6.3636 210 0.5914 0.7434 0.5914 0.7690
No log 6.4242 212 0.5877 0.7423 0.5877 0.7666
No log 6.4848 214 0.5982 0.7534 0.5982 0.7735
No log 6.5455 216 0.6268 0.7453 0.6268 0.7917
No log 6.6061 218 0.6450 0.7425 0.6450 0.8031
No log 6.6667 220 0.6369 0.7395 0.6369 0.7981
No log 6.7273 222 0.6398 0.7274 0.6398 0.7999
No log 6.7879 224 0.6476 0.7453 0.6476 0.8047
No log 6.8485 226 0.6422 0.7453 0.6422 0.8014
No log 6.9091 228 0.6232 0.7350 0.6232 0.7894
No log 6.9697 230 0.6076 0.7329 0.6076 0.7795
No log 7.0303 232 0.5938 0.7506 0.5938 0.7706
No log 7.0909 234 0.5833 0.7529 0.5833 0.7637
No log 7.1515 236 0.5699 0.7382 0.5699 0.7549
No log 7.2121 238 0.5688 0.7465 0.5688 0.7542
No log 7.2727 240 0.5768 0.7385 0.5768 0.7595
No log 7.3333 242 0.5738 0.7311 0.5738 0.7575
No log 7.3939 244 0.5697 0.7535 0.5697 0.7548
No log 7.4545 246 0.5726 0.7535 0.5726 0.7567
No log 7.5152 248 0.5760 0.7471 0.5760 0.7590
No log 7.5758 250 0.5861 0.7280 0.5861 0.7656
No log 7.6364 252 0.6072 0.7219 0.6072 0.7792
No log 7.6970 254 0.6160 0.7317 0.6160 0.7848
No log 7.7576 256 0.6193 0.7358 0.6193 0.7870
No log 7.8182 258 0.6098 0.7405 0.6098 0.7809
No log 7.8788 260 0.6000 0.7324 0.6000 0.7746
No log 7.9394 262 0.5976 0.7490 0.5976 0.7730
No log 8.0 264 0.5988 0.7490 0.5988 0.7738
No log 8.0606 266 0.5947 0.7443 0.5947 0.7712
No log 8.1212 268 0.5919 0.7490 0.5919 0.7693
No log 8.1818 270 0.5877 0.7626 0.5877 0.7666
No log 8.2424 272 0.5815 0.7506 0.5815 0.7625
No log 8.3030 274 0.5795 0.7506 0.5795 0.7613
No log 8.3636 276 0.5742 0.7465 0.5742 0.7578
No log 8.4242 278 0.5699 0.7481 0.5699 0.7549
No log 8.4848 280 0.5695 0.7454 0.5695 0.7547
No log 8.5455 282 0.5727 0.7448 0.5727 0.7568
No log 8.6061 284 0.5710 0.7454 0.5710 0.7556
No log 8.6667 286 0.5724 0.7459 0.5724 0.7566
No log 8.7273 288 0.5748 0.7481 0.5748 0.7581
No log 8.7879 290 0.5819 0.7584 0.5819 0.7628
No log 8.8485 292 0.5956 0.7708 0.5956 0.7718
No log 8.9091 294 0.6019 0.7731 0.6019 0.7758
No log 8.9697 296 0.6058 0.7677 0.6058 0.7783
No log 9.0303 298 0.5987 0.7612 0.5987 0.7738
No log 9.0909 300 0.5883 0.7649 0.5883 0.7670
No log 9.1515 302 0.5842 0.7607 0.5842 0.7643
No log 9.2121 304 0.5820 0.7487 0.5820 0.7629
No log 9.2727 306 0.5812 0.7465 0.5812 0.7624
No log 9.3333 308 0.5810 0.7459 0.5810 0.7622
No log 9.3939 310 0.5823 0.7339 0.5823 0.7631
No log 9.4545 312 0.5833 0.7339 0.5833 0.7637
No log 9.5152 314 0.5845 0.7324 0.5845 0.7645
No log 9.5758 316 0.5871 0.7465 0.5871 0.7662
No log 9.6364 318 0.5890 0.7465 0.5890 0.7675
No log 9.6970 320 0.5903 0.7465 0.5903 0.7683
No log 9.7576 322 0.5916 0.7465 0.5916 0.7692
No log 9.8182 324 0.5928 0.7465 0.5928 0.7700
No log 9.8788 326 0.5932 0.7465 0.5932 0.7702
No log 9.9394 328 0.5933 0.7465 0.5933 0.7703
No log 10.0 330 0.5932 0.7465 0.5932 0.7702

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run1_AugV5_k6_task1_organization

Finetuned
(4023)
this model