ArabicNewSplits6_FineTuningAraBERT_run1_AugV5_k4_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8322
  • Qwk: 0.5114
  • Mse: 0.8322
  • Rmse: 0.9123

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0870 2 4.0775 -0.0169 4.0775 2.0193
No log 0.1739 4 2.2392 -0.0087 2.2392 1.4964
No log 0.2609 6 1.7270 -0.0847 1.7270 1.3141
No log 0.3478 8 1.1935 0.0220 1.1935 1.0925
No log 0.4348 10 0.8457 0.0386 0.8457 0.9196
No log 0.5217 12 0.8166 0.0248 0.8166 0.9037
No log 0.6087 14 0.7296 0.1107 0.7296 0.8541
No log 0.6957 16 0.6852 0.1824 0.6852 0.8278
No log 0.7826 18 0.7200 0.1448 0.7200 0.8485
No log 0.8696 20 0.8400 0.0936 0.8400 0.9165
No log 0.9565 22 0.8663 0.0638 0.8663 0.9307
No log 1.0435 24 0.8341 0.1076 0.8341 0.9133
No log 1.1304 26 0.7658 0.2102 0.7658 0.8751
No log 1.2174 28 0.7560 0.1869 0.7560 0.8695
No log 1.3043 30 0.7793 0.1869 0.7793 0.8828
No log 1.3913 32 0.8438 0.2218 0.8438 0.9186
No log 1.4783 34 0.7536 0.2898 0.7536 0.8681
No log 1.5652 36 0.8207 0.2523 0.8207 0.9059
No log 1.6522 38 1.1425 0.1476 1.1425 1.0689
No log 1.7391 40 1.6187 0.1818 1.6187 1.2723
No log 1.8261 42 1.3389 0.2147 1.3389 1.1571
No log 1.9130 44 1.0653 0.2310 1.0653 1.0321
No log 2.0 46 1.0315 0.2494 1.0315 1.0156
No log 2.0870 48 1.1797 0.2661 1.1797 1.0862
No log 2.1739 50 1.1123 0.2892 1.1123 1.0547
No log 2.2609 52 1.0676 0.2934 1.0676 1.0332
No log 2.3478 54 0.9172 0.3807 0.9172 0.9577
No log 2.4348 56 0.7312 0.4804 0.7312 0.8551
No log 2.5217 58 0.6984 0.4916 0.6984 0.8357
No log 2.6087 60 0.6410 0.5413 0.6410 0.8006
No log 2.6957 62 0.6030 0.4545 0.6030 0.7766
No log 2.7826 64 0.6160 0.4533 0.6160 0.7848
No log 2.8696 66 0.7119 0.4272 0.7119 0.8438
No log 2.9565 68 0.7127 0.4272 0.7127 0.8442
No log 3.0435 70 0.6121 0.4319 0.6121 0.7823
No log 3.1304 72 0.5738 0.4538 0.5738 0.7575
No log 3.2174 74 0.5792 0.4909 0.5792 0.7611
No log 3.3043 76 0.6109 0.5128 0.6109 0.7816
No log 3.3913 78 0.6238 0.5456 0.6238 0.7898
No log 3.4783 80 0.6617 0.5672 0.6617 0.8134
No log 3.5652 82 0.6850 0.5476 0.6850 0.8276
No log 3.6522 84 0.7079 0.5663 0.7079 0.8414
No log 3.7391 86 0.7441 0.5662 0.7441 0.8626
No log 3.8261 88 0.7521 0.5294 0.7521 0.8672
No log 3.9130 90 0.7590 0.4710 0.7590 0.8712
No log 4.0 92 0.7365 0.5294 0.7365 0.8582
No log 4.0870 94 0.7242 0.5478 0.7242 0.8510
No log 4.1739 96 0.7499 0.5760 0.7499 0.8659
No log 4.2609 98 0.7452 0.5763 0.7452 0.8633
No log 4.3478 100 0.7145 0.5356 0.7145 0.8453
No log 4.4348 102 0.7472 0.4294 0.7472 0.8644
No log 4.5217 104 0.7922 0.3968 0.7922 0.8900
No log 4.6087 106 0.7673 0.3958 0.7673 0.8760
No log 4.6957 108 0.7369 0.4831 0.7369 0.8584
No log 4.7826 110 0.7603 0.5748 0.7603 0.8719
No log 4.8696 112 0.7912 0.5415 0.7912 0.8895
No log 4.9565 114 0.7940 0.5385 0.7940 0.8911
No log 5.0435 116 0.8002 0.4548 0.8002 0.8945
No log 5.1304 118 0.8612 0.4202 0.8612 0.9280
No log 5.2174 120 0.8737 0.4235 0.8737 0.9347
No log 5.3043 122 0.7996 0.4272 0.7996 0.8942
No log 5.3913 124 0.7523 0.4845 0.7523 0.8673
No log 5.4783 126 0.7666 0.5333 0.7666 0.8755
No log 5.5652 128 0.7873 0.5696 0.7873 0.8873
No log 5.6522 130 0.8185 0.5624 0.8185 0.9047
No log 5.7391 132 0.8080 0.5178 0.8080 0.8989
No log 5.8261 134 0.8222 0.5245 0.8222 0.9068
No log 5.9130 136 0.8485 0.5117 0.8485 0.9212
No log 6.0 138 0.8673 0.5254 0.8673 0.9313
No log 6.0870 140 0.8830 0.5337 0.8830 0.9397
No log 6.1739 142 0.8907 0.5460 0.8907 0.9438
No log 6.2609 144 0.8890 0.5566 0.8890 0.9429
No log 6.3478 146 0.8791 0.5460 0.8791 0.9376
No log 6.4348 148 0.8696 0.5376 0.8696 0.9325
No log 6.5217 150 0.8663 0.5376 0.8663 0.9308
No log 6.6087 152 0.8609 0.5249 0.8609 0.9278
No log 6.6957 154 0.8697 0.5421 0.8697 0.9326
No log 6.7826 156 0.8790 0.5369 0.8790 0.9376
No log 6.8696 158 0.8813 0.5447 0.8813 0.9388
No log 6.9565 160 0.8858 0.5487 0.8858 0.9412
No log 7.0435 162 0.8795 0.5353 0.8795 0.9378
No log 7.1304 164 0.8819 0.54 0.8819 0.9391
No log 7.2174 166 0.8807 0.5438 0.8807 0.9385
No log 7.3043 168 0.8616 0.5345 0.8616 0.9282
No log 7.3913 170 0.8595 0.5312 0.8595 0.9271
No log 7.4783 172 0.8493 0.5248 0.8493 0.9216
No log 7.5652 174 0.8554 0.5461 0.8554 0.9249
No log 7.6522 176 0.8568 0.5461 0.8568 0.9256
No log 7.7391 178 0.8545 0.5461 0.8545 0.9244
No log 7.8261 180 0.8563 0.5473 0.8563 0.9254
No log 7.9130 182 0.8567 0.5381 0.8567 0.9256
No log 8.0 184 0.8627 0.5370 0.8627 0.9288
No log 8.0870 186 0.8736 0.5370 0.8736 0.9346
No log 8.1739 188 0.8867 0.5474 0.8867 0.9417
No log 8.2609 190 0.8879 0.5435 0.8879 0.9423
No log 8.3478 192 0.8872 0.5364 0.8872 0.9419
No log 8.4348 194 0.8833 0.5103 0.8833 0.9399
No log 8.5217 196 0.8798 0.5020 0.8798 0.9380
No log 8.6087 198 0.8721 0.5103 0.8721 0.9339
No log 8.6957 200 0.8658 0.5364 0.8658 0.9305
No log 8.7826 202 0.8604 0.5483 0.8604 0.9276
No log 8.8696 204 0.8553 0.5322 0.8553 0.9248
No log 8.9565 206 0.8598 0.5252 0.8598 0.9272
No log 9.0435 208 0.8543 0.5256 0.8543 0.9243
No log 9.1304 210 0.8451 0.5553 0.8451 0.9193
No log 9.2174 212 0.8369 0.5319 0.8369 0.9148
No log 9.3043 214 0.8294 0.5304 0.8294 0.9107
No log 9.3913 216 0.8260 0.5346 0.8260 0.9089
No log 9.4783 218 0.8264 0.5219 0.8264 0.9091
No log 9.5652 220 0.8293 0.5123 0.8293 0.9107
No log 9.6522 222 0.8306 0.5123 0.8306 0.9114
No log 9.7391 224 0.8310 0.5130 0.8310 0.9116
No log 9.8261 226 0.8314 0.5130 0.8314 0.9118
No log 9.9130 228 0.8319 0.5114 0.8319 0.9121
No log 10.0 230 0.8322 0.5114 0.8322 0.9123

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run1_AugV5_k4_task2_organization

Finetuned
(4023)
this model