ArabicNewSplits6_FineTuningAraBERT_run3_AugV5_k11_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6512
  • Qwk: 0.7261
  • Mse: 0.6512
  • Rmse: 0.8070

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0488 2 2.4036 0.0144 2.4036 1.5503
No log 0.0976 4 1.7401 0.1026 1.7401 1.3191
No log 0.1463 6 1.6614 0.0583 1.6614 1.2890
No log 0.1951 8 1.5960 0.1224 1.5960 1.2633
No log 0.2439 10 1.4494 0.1796 1.4494 1.2039
No log 0.2927 12 1.3154 0.1764 1.3154 1.1469
No log 0.3415 14 1.2657 0.1601 1.2657 1.1250
No log 0.3902 16 1.2312 0.1842 1.2312 1.1096
No log 0.4390 18 1.1792 0.2119 1.1792 1.0859
No log 0.4878 20 1.1570 0.2355 1.1570 1.0756
No log 0.5366 22 1.2329 0.3789 1.2329 1.1104
No log 0.5854 24 1.2167 0.4190 1.2167 1.1031
No log 0.6341 26 1.1021 0.3964 1.1021 1.0498
No log 0.6829 28 1.0330 0.3731 1.0330 1.0164
No log 0.7317 30 1.0070 0.4080 1.0070 1.0035
No log 0.7805 32 0.9791 0.4396 0.9791 0.9895
No log 0.8293 34 1.0022 0.5156 1.0022 1.0011
No log 0.8780 36 1.0271 0.5305 1.0271 1.0135
No log 0.9268 38 1.0051 0.5305 1.0051 1.0025
No log 0.9756 40 0.9250 0.5097 0.9250 0.9617
No log 1.0244 42 0.8761 0.5143 0.8761 0.9360
No log 1.0732 44 0.9079 0.4938 0.9079 0.9528
No log 1.1220 46 0.8852 0.5355 0.8852 0.9409
No log 1.1707 48 0.8704 0.5140 0.8704 0.9330
No log 1.2195 50 1.1465 0.4955 1.1465 1.0708
No log 1.2683 52 1.1463 0.5046 1.1463 1.0707
No log 1.3171 54 0.9686 0.5153 0.9686 0.9842
No log 1.3659 56 1.0092 0.5009 1.0092 1.0046
No log 1.4146 58 0.9839 0.5078 0.9839 0.9919
No log 1.4634 60 0.9616 0.5052 0.9616 0.9806
No log 1.5122 62 0.9559 0.5039 0.9559 0.9777
No log 1.5610 64 0.9865 0.5329 0.9865 0.9933
No log 1.6098 66 1.0114 0.5448 1.0114 1.0057
No log 1.6585 68 0.9515 0.5617 0.9515 0.9755
No log 1.7073 70 0.9306 0.5567 0.9306 0.9647
No log 1.7561 72 0.7878 0.6364 0.7878 0.8876
No log 1.8049 74 0.7944 0.5869 0.7944 0.8913
No log 1.8537 76 0.8223 0.5752 0.8223 0.9068
No log 1.9024 78 0.7889 0.6255 0.7889 0.8882
No log 1.9512 80 0.8189 0.5620 0.8189 0.9049
No log 2.0 82 0.9124 0.5257 0.9124 0.9552
No log 2.0488 84 1.0048 0.5638 1.0048 1.0024
No log 2.0976 86 0.9457 0.5567 0.9457 0.9725
No log 2.1463 88 0.8679 0.6101 0.8679 0.9316
No log 2.1951 90 0.7590 0.6208 0.7590 0.8712
No log 2.2439 92 0.7039 0.6549 0.7039 0.8390
No log 2.2927 94 0.6865 0.6772 0.6865 0.8286
No log 2.3415 96 0.6783 0.6876 0.6783 0.8236
No log 2.3902 98 0.6886 0.6931 0.6886 0.8298
No log 2.4390 100 0.7032 0.7101 0.7032 0.8386
No log 2.4878 102 0.7051 0.7015 0.7051 0.8397
No log 2.5366 104 0.7134 0.7116 0.7134 0.8447
No log 2.5854 106 0.6968 0.6954 0.6968 0.8347
No log 2.6341 108 0.7022 0.6903 0.7022 0.8380
No log 2.6829 110 0.7168 0.6885 0.7168 0.8466
No log 2.7317 112 0.7479 0.7004 0.7479 0.8648
No log 2.7805 114 0.7666 0.6526 0.7666 0.8756
No log 2.8293 116 0.7647 0.6785 0.7647 0.8745
No log 2.8780 118 0.7362 0.6895 0.7362 0.8580
No log 2.9268 120 0.6799 0.6891 0.6799 0.8245
No log 2.9756 122 0.6808 0.6600 0.6808 0.8251
No log 3.0244 124 0.6852 0.6270 0.6852 0.8278
No log 3.0732 126 0.6959 0.6716 0.6959 0.8342
No log 3.1220 128 0.7188 0.6789 0.7188 0.8478
No log 3.1707 130 0.7451 0.6908 0.7451 0.8632
No log 3.2195 132 0.8230 0.7050 0.8230 0.9072
No log 3.2683 134 0.7590 0.7295 0.7590 0.8712
No log 3.3171 136 0.6708 0.7308 0.6708 0.8190
No log 3.3659 138 0.6433 0.7192 0.6433 0.8021
No log 3.4146 140 0.6441 0.7190 0.6441 0.8025
No log 3.4634 142 0.7459 0.7393 0.7459 0.8637
No log 3.5122 144 0.8158 0.7296 0.8158 0.9032
No log 3.5610 146 0.7074 0.7272 0.7074 0.8410
No log 3.6098 148 0.6098 0.7276 0.6098 0.7809
No log 3.6585 150 0.5709 0.7126 0.5709 0.7556
No log 3.7073 152 0.5717 0.7026 0.5717 0.7561
No log 3.7561 154 0.5647 0.7010 0.5647 0.7515
No log 3.8049 156 0.6560 0.7063 0.6560 0.8099
No log 3.8537 158 0.7709 0.7226 0.7709 0.8780
No log 3.9024 160 0.7765 0.7217 0.7765 0.8812
No log 3.9512 162 0.6569 0.7188 0.6569 0.8105
No log 4.0 164 0.5871 0.7186 0.5871 0.7662
No log 4.0488 166 0.5819 0.6825 0.5819 0.7628
No log 4.0976 168 0.5853 0.6994 0.5853 0.7651
No log 4.1463 170 0.6252 0.7205 0.6252 0.7907
No log 4.1951 172 0.7329 0.6904 0.7329 0.8561
No log 4.2439 174 0.7523 0.6904 0.7523 0.8674
No log 4.2927 176 0.6862 0.7173 0.6862 0.8283
No log 4.3415 178 0.5988 0.7049 0.5988 0.7738
No log 4.3902 180 0.5871 0.6865 0.5871 0.7662
No log 4.4390 182 0.6013 0.6897 0.6013 0.7754
No log 4.4878 184 0.5894 0.6949 0.5894 0.7677
No log 4.5366 186 0.5725 0.7066 0.5725 0.7566
No log 4.5854 188 0.6413 0.7555 0.6413 0.8008
No log 4.6341 190 0.8267 0.6961 0.8267 0.9092
No log 4.6829 192 0.9010 0.6654 0.9010 0.9492
No log 4.7317 194 0.8300 0.6864 0.8300 0.9110
No log 4.7805 196 0.7231 0.7435 0.7231 0.8504
No log 4.8293 198 0.6284 0.7595 0.6284 0.7927
No log 4.8780 200 0.5955 0.7291 0.5955 0.7717
No log 4.9268 202 0.6048 0.7417 0.6048 0.7777
No log 4.9756 204 0.6117 0.7417 0.6117 0.7821
No log 5.0244 206 0.6257 0.7638 0.6257 0.7910
No log 5.0732 208 0.6599 0.7474 0.6599 0.8123
No log 5.1220 210 0.7141 0.7301 0.7141 0.8450
No log 5.1707 212 0.7100 0.7301 0.7100 0.8426
No log 5.2195 214 0.6791 0.7293 0.6791 0.8241
No log 5.2683 216 0.6354 0.7417 0.6354 0.7971
No log 5.3171 218 0.6097 0.6953 0.6097 0.7808
No log 5.3659 220 0.6011 0.7001 0.6011 0.7753
No log 5.4146 222 0.5988 0.7001 0.5988 0.7738
No log 5.4634 224 0.5992 0.7045 0.5992 0.7741
No log 5.5122 226 0.6185 0.7164 0.6185 0.7865
No log 5.5610 228 0.6813 0.7107 0.6813 0.8254
No log 5.6098 230 0.7367 0.7013 0.7367 0.8583
No log 5.6585 232 0.7275 0.7013 0.7275 0.8529
No log 5.7073 234 0.7066 0.7054 0.7066 0.8406
No log 5.7561 236 0.6634 0.6930 0.6634 0.8145
No log 5.8049 238 0.6508 0.6926 0.6508 0.8067
No log 5.8537 240 0.6422 0.6972 0.6422 0.8014
No log 5.9024 242 0.6362 0.7208 0.6362 0.7976
No log 5.9512 244 0.6192 0.7454 0.6192 0.7869
No log 6.0 246 0.6140 0.7330 0.6140 0.7836
No log 6.0488 248 0.6130 0.7330 0.6130 0.7829
No log 6.0976 250 0.6068 0.7376 0.6068 0.7790
No log 6.1463 252 0.6101 0.7307 0.6101 0.7811
No log 6.1951 254 0.6071 0.7353 0.6071 0.7792
No log 6.2439 256 0.5964 0.7291 0.5964 0.7723
No log 6.2927 258 0.5930 0.7204 0.5930 0.7701
No log 6.3415 260 0.5987 0.7204 0.5987 0.7737
No log 6.3902 262 0.5942 0.7204 0.5942 0.7709
No log 6.4390 264 0.5883 0.7164 0.5883 0.7670
No log 6.4878 266 0.5839 0.7170 0.5839 0.7641
No log 6.5366 268 0.5893 0.7164 0.5893 0.7677
No log 6.5854 270 0.6071 0.7307 0.6071 0.7792
No log 6.6341 272 0.6397 0.7518 0.6397 0.7998
No log 6.6829 274 0.6508 0.7357 0.6508 0.8067
No log 6.7317 276 0.6534 0.7400 0.6534 0.8084
No log 6.7805 278 0.6511 0.7476 0.6511 0.8069
No log 6.8293 280 0.6341 0.7277 0.6341 0.7963
No log 6.8780 282 0.6157 0.7480 0.6157 0.7846
No log 6.9268 284 0.6112 0.7441 0.6112 0.7818
No log 6.9756 286 0.6134 0.7441 0.6134 0.7832
No log 7.0244 288 0.6207 0.7277 0.6207 0.7879
No log 7.0732 290 0.6354 0.7357 0.6354 0.7971
No log 7.1220 292 0.6424 0.7357 0.6424 0.8015
No log 7.1707 294 0.6596 0.7293 0.6596 0.8122
No log 7.2195 296 0.6507 0.7417 0.6507 0.8067
No log 7.2683 298 0.6221 0.7493 0.6221 0.7887
No log 7.3171 300 0.5938 0.7228 0.5938 0.7706
No log 7.3659 302 0.5843 0.7067 0.5843 0.7644
No log 7.4146 304 0.5874 0.7067 0.5874 0.7664
No log 7.4634 306 0.6031 0.7149 0.6031 0.7766
No log 7.5122 308 0.6289 0.7369 0.6289 0.7930
No log 7.5610 310 0.6305 0.7451 0.6305 0.7940
No log 7.6098 312 0.6399 0.7344 0.6399 0.8000
No log 7.6585 314 0.6390 0.7344 0.6390 0.7994
No log 7.7073 316 0.6434 0.7344 0.6434 0.8021
No log 7.7561 318 0.6554 0.7440 0.6554 0.8095
No log 7.8049 320 0.6592 0.7440 0.6592 0.8119
No log 7.8537 322 0.6709 0.7478 0.6709 0.8191
No log 7.9024 324 0.6817 0.7191 0.6817 0.8257
No log 7.9512 326 0.6748 0.7234 0.6748 0.8214
No log 8.0 328 0.6671 0.7234 0.6671 0.8168
No log 8.0488 330 0.6586 0.7322 0.6586 0.8116
No log 8.0976 332 0.6524 0.7262 0.6524 0.8077
No log 8.1463 334 0.6575 0.7322 0.6575 0.8109
No log 8.1951 336 0.6627 0.7196 0.6627 0.8140
No log 8.2439 338 0.6649 0.7110 0.6649 0.8154
No log 8.2927 340 0.6546 0.7199 0.6546 0.8091
No log 8.3415 342 0.6517 0.7199 0.6517 0.8073
No log 8.3902 344 0.6588 0.7384 0.6588 0.8116
No log 8.4390 346 0.6620 0.7339 0.6620 0.8137
No log 8.4878 348 0.6576 0.7338 0.6576 0.8109
No log 8.5366 350 0.6474 0.7307 0.6474 0.8046
No log 8.5854 352 0.6432 0.7307 0.6432 0.8020
No log 8.6341 354 0.6434 0.7353 0.6434 0.8021
No log 8.6829 356 0.6464 0.7514 0.6464 0.8040
No log 8.7317 358 0.6465 0.7328 0.6465 0.8040
No log 8.7805 360 0.6425 0.7247 0.6425 0.8016
No log 8.8293 362 0.6438 0.7247 0.6438 0.8024
No log 8.8780 364 0.6511 0.7387 0.6511 0.8069
No log 8.9268 366 0.6591 0.7364 0.6591 0.8118
No log 8.9756 368 0.6634 0.7518 0.6634 0.8145
No log 9.0244 370 0.6692 0.7395 0.6692 0.8181
No log 9.0732 372 0.6753 0.7388 0.6753 0.8217
No log 9.1220 374 0.6791 0.7388 0.6791 0.8241
No log 9.1707 376 0.6790 0.7388 0.6790 0.8240
No log 9.2195 378 0.6797 0.7388 0.6797 0.8245
No log 9.2683 380 0.6853 0.7388 0.6853 0.8278
No log 9.3171 382 0.6922 0.7228 0.6922 0.8320
No log 9.3659 384 0.6948 0.7185 0.6948 0.8336
No log 9.4146 386 0.6942 0.7185 0.6942 0.8332
No log 9.4634 388 0.6887 0.7345 0.6887 0.8299
No log 9.5122 390 0.6846 0.7388 0.6846 0.8274
No log 9.5610 392 0.6777 0.7388 0.6777 0.8232
No log 9.6098 394 0.6728 0.7388 0.6728 0.8202
No log 9.6585 396 0.6687 0.7388 0.6687 0.8177
No log 9.7073 398 0.6642 0.7512 0.6642 0.8150
No log 9.7561 400 0.6595 0.7316 0.6595 0.8121
No log 9.8049 402 0.6557 0.7277 0.6557 0.8097
No log 9.8537 404 0.6536 0.7300 0.6536 0.8084
No log 9.9024 406 0.6523 0.7261 0.6523 0.8076
No log 9.9512 408 0.6513 0.7261 0.6513 0.8070
No log 10.0 410 0.6512 0.7261 0.6512 0.8070

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run3_AugV5_k11_task5_organization

Finetuned
(4023)
this model