ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k12_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7608
  • Qwk: 0.5987
  • Mse: 0.7608
  • Rmse: 0.8722

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0333 2 5.4299 -0.0309 5.4299 2.3302
No log 0.0667 4 3.3388 0.0464 3.3388 1.8272
No log 0.1 6 2.0115 0.0763 2.0115 1.4183
No log 0.1333 8 1.6188 0.1013 1.6188 1.2723
No log 0.1667 10 1.4419 0.0860 1.4419 1.2008
No log 0.2 12 1.2368 0.2115 1.2368 1.1121
No log 0.2333 14 1.2231 0.1947 1.2231 1.1059
No log 0.2667 16 1.5365 -0.0185 1.5365 1.2396
No log 0.3 18 1.8738 0.0348 1.8738 1.3689
No log 0.3333 20 2.1152 0.1244 2.1152 1.4544
No log 0.3667 22 1.8170 0.0376 1.8170 1.3480
No log 0.4 24 1.6215 0.0190 1.6215 1.2734
No log 0.4333 26 1.3853 0.0812 1.3853 1.1770
No log 0.4667 28 1.2328 0.3505 1.2328 1.1103
No log 0.5 30 1.1520 0.3293 1.1520 1.0733
No log 0.5333 32 1.1071 0.3443 1.1071 1.0522
No log 0.5667 34 1.0876 0.3205 1.0876 1.0429
No log 0.6 36 1.1276 0.3528 1.1276 1.0619
No log 0.6333 38 1.1987 0.3124 1.1987 1.0949
No log 0.6667 40 1.5257 0.1200 1.5257 1.2352
No log 0.7 42 1.6329 0.1233 1.6329 1.2779
No log 0.7333 44 1.7263 0.1147 1.7263 1.3139
No log 0.7667 46 1.4035 0.1741 1.4035 1.1847
No log 0.8 48 1.3953 0.1723 1.3953 1.1812
No log 0.8333 50 1.7092 0.1852 1.7092 1.3074
No log 0.8667 52 2.0678 0.2137 2.0678 1.4380
No log 0.9 54 1.6432 0.2187 1.6432 1.2819
No log 0.9333 56 1.2670 0.2475 1.2670 1.1256
No log 0.9667 58 1.0557 0.4505 1.0557 1.0275
No log 1.0 60 1.0431 0.4644 1.0431 1.0213
No log 1.0333 62 1.0890 0.4559 1.0890 1.0436
No log 1.0667 64 1.4979 0.3147 1.4979 1.2239
No log 1.1 66 1.7355 0.2894 1.7355 1.3174
No log 1.1333 68 1.4408 0.3674 1.4408 1.2003
No log 1.1667 70 0.9154 0.5132 0.9154 0.9568
No log 1.2 72 0.6814 0.5763 0.6814 0.8255
No log 1.2333 74 0.6978 0.5799 0.6978 0.8354
No log 1.2667 76 0.6706 0.6279 0.6706 0.8189
No log 1.3 78 0.7918 0.5601 0.7918 0.8898
No log 1.3333 80 1.1236 0.4243 1.1236 1.0600
No log 1.3667 82 1.2134 0.4020 1.2134 1.1015
No log 1.4 84 1.0613 0.4831 1.0613 1.0302
No log 1.4333 86 0.8485 0.5620 0.8485 0.9211
No log 1.4667 88 0.7656 0.6060 0.7656 0.8750
No log 1.5 90 0.9166 0.5467 0.9166 0.9574
No log 1.5333 92 1.3719 0.4696 1.3719 1.1713
No log 1.5667 94 1.4879 0.4571 1.4879 1.2198
No log 1.6 96 1.2028 0.5493 1.2028 1.0967
No log 1.6333 98 0.7931 0.6784 0.7931 0.8906
No log 1.6667 100 0.6655 0.7042 0.6655 0.8158
No log 1.7 102 0.6370 0.7117 0.6370 0.7981
No log 1.7333 104 0.6321 0.7372 0.6321 0.7950
No log 1.7667 106 0.6273 0.7168 0.6273 0.7920
No log 1.8 108 0.5951 0.7507 0.5951 0.7714
No log 1.8333 110 0.6405 0.6936 0.6405 0.8003
No log 1.8667 112 0.6428 0.6885 0.6428 0.8017
No log 1.9 114 0.6722 0.6939 0.6722 0.8199
No log 1.9333 116 0.8233 0.5704 0.8233 0.9074
No log 1.9667 118 1.0099 0.5426 1.0099 1.0049
No log 2.0 120 0.8867 0.5575 0.8867 0.9417
No log 2.0333 122 0.6481 0.6778 0.6481 0.8051
No log 2.0667 124 0.5995 0.7004 0.5995 0.7743
No log 2.1 126 0.6052 0.7077 0.6052 0.7779
No log 2.1333 128 0.6147 0.7356 0.6147 0.7840
No log 2.1667 130 0.6283 0.7224 0.6283 0.7927
No log 2.2 132 0.7082 0.6672 0.7082 0.8415
No log 2.2333 134 1.0197 0.5431 1.0197 1.0098
No log 2.2667 136 1.1403 0.5163 1.1403 1.0679
No log 2.3 138 1.0130 0.5361 1.0130 1.0065
No log 2.3333 140 0.6468 0.7019 0.6468 0.8042
No log 2.3667 142 0.5874 0.7410 0.5874 0.7664
No log 2.4 144 0.5976 0.7147 0.5976 0.7731
No log 2.4333 146 0.5955 0.7328 0.5955 0.7717
No log 2.4667 148 0.6051 0.7291 0.6051 0.7779
No log 2.5 150 0.6274 0.7386 0.6274 0.7921
No log 2.5333 152 0.6071 0.7250 0.6071 0.7792
No log 2.5667 154 0.7206 0.6122 0.7206 0.8489
No log 2.6 156 0.6818 0.6479 0.6818 0.8257
No log 2.6333 158 0.6211 0.7276 0.6211 0.7881
No log 2.6667 160 0.7733 0.6161 0.7733 0.8794
No log 2.7 162 0.7663 0.6244 0.7663 0.8754
No log 2.7333 164 0.6352 0.7251 0.6352 0.7970
No log 2.7667 166 0.7462 0.6310 0.7462 0.8638
No log 2.8 168 0.9689 0.5774 0.9689 0.9843
No log 2.8333 170 0.8994 0.5776 0.8994 0.9483
No log 2.8667 172 0.7153 0.6483 0.7153 0.8457
No log 2.9 174 0.7020 0.6623 0.7020 0.8379
No log 2.9333 176 0.7482 0.6487 0.7482 0.8650
No log 2.9667 178 0.7105 0.6621 0.7105 0.8429
No log 3.0 180 0.7270 0.6143 0.7270 0.8527
No log 3.0333 182 0.7704 0.5731 0.7704 0.8777
No log 3.0667 184 0.6990 0.6443 0.6990 0.8361
No log 3.1 186 0.6181 0.6615 0.6181 0.7862
No log 3.1333 188 0.6083 0.6863 0.6083 0.7800
No log 3.1667 190 0.5950 0.7010 0.5950 0.7714
No log 3.2 192 0.6107 0.7000 0.6107 0.7815
No log 3.2333 194 0.6724 0.6744 0.6724 0.8200
No log 3.2667 196 0.6766 0.6604 0.6766 0.8226
No log 3.3 198 0.6120 0.6907 0.6120 0.7823
No log 3.3333 200 0.6220 0.7177 0.6220 0.7887
No log 3.3667 202 0.6317 0.7109 0.6317 0.7948
No log 3.4 204 0.6150 0.7029 0.6150 0.7842
No log 3.4333 206 0.7712 0.5733 0.7712 0.8782
No log 3.4667 208 0.8293 0.5650 0.8293 0.9106
No log 3.5 210 0.7066 0.6799 0.7066 0.8406
No log 3.5333 212 0.6326 0.7273 0.6326 0.7954
No log 3.5667 214 0.6394 0.7374 0.6394 0.7996
No log 3.6 216 0.7082 0.7073 0.7082 0.8415
No log 3.6333 218 0.7115 0.7202 0.7115 0.8435
No log 3.6667 220 0.7339 0.7013 0.7339 0.8567
No log 3.7 222 0.6851 0.7154 0.6851 0.8277
No log 3.7333 224 0.6262 0.7259 0.6262 0.7914
No log 3.7667 226 0.6664 0.6943 0.6664 0.8163
No log 3.8 228 0.6610 0.6882 0.6610 0.8130
No log 3.8333 230 0.6226 0.7058 0.6226 0.7890
No log 3.8667 232 0.6557 0.6347 0.6557 0.8098
No log 3.9 234 0.6449 0.6454 0.6449 0.8031
No log 3.9333 236 0.6271 0.7077 0.6271 0.7919
No log 3.9667 238 0.6598 0.7110 0.6598 0.8123
No log 4.0 240 0.7456 0.7126 0.7456 0.8635
No log 4.0333 242 0.8129 0.6883 0.8129 0.9016
No log 4.0667 244 0.7281 0.6954 0.7281 0.8533
No log 4.1 246 0.6191 0.7136 0.6191 0.7868
No log 4.1333 248 0.6237 0.6982 0.6237 0.7898
No log 4.1667 250 0.6402 0.6596 0.6402 0.8002
No log 4.2 252 0.6144 0.6885 0.6144 0.7839
No log 4.2333 254 0.6496 0.6827 0.6496 0.8060
No log 4.2667 256 0.6814 0.6654 0.6814 0.8255
No log 4.3 258 0.6479 0.6575 0.6479 0.8049
No log 4.3333 260 0.6205 0.7143 0.6205 0.7877
No log 4.3667 262 0.6068 0.7488 0.6068 0.7790
No log 4.4 264 0.6437 0.6694 0.6437 0.8023
No log 4.4333 266 0.6740 0.6760 0.6740 0.8210
No log 4.4667 268 0.5939 0.7091 0.5939 0.7707
No log 4.5 270 0.5793 0.7447 0.5793 0.7611
No log 4.5333 272 0.5727 0.7224 0.5727 0.7568
No log 4.5667 274 0.5632 0.7187 0.5632 0.7505
No log 4.6 276 0.5722 0.7255 0.5722 0.7565
No log 4.6333 278 0.5529 0.7196 0.5529 0.7436
No log 4.6667 280 0.5569 0.7337 0.5569 0.7463
No log 4.7 282 0.5547 0.7267 0.5547 0.7448
No log 4.7333 284 0.5893 0.6972 0.5893 0.7677
No log 4.7667 286 0.5968 0.6972 0.5968 0.7725
No log 4.8 288 0.6146 0.6831 0.6146 0.7840
No log 4.8333 290 0.5835 0.7105 0.5835 0.7639
No log 4.8667 292 0.5801 0.7203 0.5801 0.7617
No log 4.9 294 0.6032 0.6896 0.6032 0.7767
No log 4.9333 296 0.6080 0.6932 0.6080 0.7798
No log 4.9667 298 0.6037 0.6929 0.6037 0.7770
No log 5.0 300 0.6017 0.6598 0.6017 0.7757
No log 5.0333 302 0.6175 0.6714 0.6175 0.7858
No log 5.0667 304 0.6187 0.6694 0.6187 0.7865
No log 5.1 306 0.6277 0.6674 0.6277 0.7923
No log 5.1333 308 0.6330 0.6215 0.6330 0.7956
No log 5.1667 310 0.7368 0.6134 0.7368 0.8584
No log 5.2 312 0.7651 0.6055 0.7651 0.8747
No log 5.2333 314 0.6639 0.6365 0.6639 0.8148
No log 5.2667 316 0.6196 0.6694 0.6196 0.7871
No log 5.3 318 0.7504 0.6318 0.7504 0.8662
No log 5.3333 320 0.8052 0.6582 0.8052 0.8973
No log 5.3667 322 0.7032 0.6587 0.7032 0.8386
No log 5.4 324 0.6076 0.7082 0.6076 0.7795
No log 5.4333 326 0.6359 0.6968 0.6359 0.7974
No log 5.4667 328 0.6198 0.7169 0.6198 0.7873
No log 5.5 330 0.5993 0.7017 0.5993 0.7741
No log 5.5333 332 0.6629 0.6806 0.6629 0.8142
No log 5.5667 334 0.6776 0.6869 0.6776 0.8232
No log 5.6 336 0.6261 0.6800 0.6261 0.7913
No log 5.6333 338 0.6051 0.6766 0.6051 0.7779
No log 5.6667 340 0.5950 0.6781 0.5950 0.7714
No log 5.7 342 0.6072 0.6849 0.6072 0.7792
No log 5.7333 344 0.6252 0.6816 0.6252 0.7907
No log 5.7667 346 0.7082 0.6693 0.7082 0.8415
No log 5.8 348 0.7172 0.6498 0.7172 0.8469
No log 5.8333 350 0.6705 0.6511 0.6705 0.8189
No log 5.8667 352 0.6105 0.6850 0.6105 0.7813
No log 5.9 354 0.6032 0.6693 0.6032 0.7767
No log 5.9333 356 0.5941 0.6858 0.5941 0.7708
No log 5.9667 358 0.5966 0.7409 0.5966 0.7724
No log 6.0 360 0.6101 0.7346 0.6101 0.7811
No log 6.0333 362 0.6251 0.7122 0.6251 0.7907
No log 6.0667 364 0.6131 0.7291 0.6131 0.7830
No log 6.1 366 0.6005 0.7466 0.6005 0.7749
No log 6.1333 368 0.6015 0.7109 0.6015 0.7756
No log 6.1667 370 0.6239 0.6977 0.6239 0.7899
No log 6.2 372 0.6288 0.6948 0.6288 0.7930
No log 6.2333 374 0.6382 0.6888 0.6382 0.7989
No log 6.2667 376 0.6177 0.7039 0.6177 0.7860
No log 6.3 378 0.5786 0.72 0.5786 0.7607
No log 6.3333 380 0.5675 0.7157 0.5675 0.7533
No log 6.3667 382 0.5718 0.7172 0.5718 0.7562
No log 6.4 384 0.6538 0.7106 0.6538 0.8086
No log 6.4333 386 0.7045 0.6876 0.7045 0.8394
No log 6.4667 388 0.6259 0.7212 0.6259 0.7912
No log 6.5 390 0.5715 0.7233 0.5715 0.7560
No log 6.5333 392 0.6080 0.6847 0.6080 0.7797
No log 6.5667 394 0.6238 0.6932 0.6238 0.7898
No log 6.6 396 0.5972 0.7244 0.5972 0.7728
No log 6.6333 398 0.5937 0.7291 0.5937 0.7705
No log 6.6667 400 0.5974 0.7329 0.5974 0.7729
No log 6.7 402 0.6138 0.7045 0.6138 0.7834
No log 6.7333 404 0.6588 0.6786 0.6588 0.8117
No log 6.7667 406 0.7235 0.6111 0.7235 0.8506
No log 6.8 408 0.6765 0.6537 0.6765 0.8225
No log 6.8333 410 0.6017 0.6620 0.6017 0.7757
No log 6.8667 412 0.5921 0.7132 0.5921 0.7695
No log 6.9 414 0.6388 0.6884 0.6388 0.7993
No log 6.9333 416 0.6203 0.7016 0.6203 0.7876
No log 6.9667 418 0.5847 0.7608 0.5847 0.7647
No log 7.0 420 0.6511 0.6778 0.6511 0.8069
No log 7.0333 422 0.6902 0.6477 0.6902 0.8308
No log 7.0667 424 0.6493 0.6809 0.6493 0.8058
No log 7.1 426 0.6045 0.7165 0.6045 0.7775
No log 7.1333 428 0.5977 0.7378 0.5977 0.7731
No log 7.1667 430 0.6004 0.7194 0.6004 0.7748
No log 7.2 432 0.6552 0.6624 0.6552 0.8094
No log 7.2333 434 0.7776 0.6341 0.7776 0.8818
No log 7.2667 436 0.7945 0.6201 0.7945 0.8913
No log 7.3 438 0.7243 0.6432 0.7243 0.8511
No log 7.3333 440 0.6238 0.6790 0.6238 0.7898
No log 7.3667 442 0.5894 0.6807 0.5894 0.7677
No log 7.4 444 0.6362 0.7014 0.6362 0.7976
No log 7.4333 446 0.6337 0.7128 0.6337 0.7961
No log 7.4667 448 0.6005 0.6613 0.6005 0.7749
No log 7.5 450 0.6102 0.6770 0.6102 0.7812
No log 7.5333 452 0.6235 0.6842 0.6235 0.7896
No log 7.5667 454 0.5949 0.6808 0.5949 0.7713
No log 7.6 456 0.6009 0.6925 0.6009 0.7752
No log 7.6333 458 0.6359 0.7151 0.6359 0.7974
No log 7.6667 460 0.6469 0.7147 0.6469 0.8043
No log 7.7 462 0.6038 0.7182 0.6038 0.7770
No log 7.7333 464 0.6112 0.6914 0.6112 0.7818
No log 7.7667 466 0.6675 0.6695 0.6675 0.8170
No log 7.8 468 0.7424 0.6507 0.7424 0.8617
No log 7.8333 470 0.7117 0.6768 0.7117 0.8436
No log 7.8667 472 0.6584 0.6644 0.6584 0.8114
No log 7.9 474 0.6656 0.6744 0.6656 0.8158
No log 7.9333 476 0.7174 0.6896 0.7174 0.8470
No log 7.9667 478 0.6977 0.6961 0.6977 0.8353
No log 8.0 480 0.6375 0.6843 0.6375 0.7985
No log 8.0333 482 0.6081 0.7031 0.6081 0.7798
No log 8.0667 484 0.6035 0.6952 0.6035 0.7768
No log 8.1 486 0.5991 0.7316 0.5991 0.7740
No log 8.1333 488 0.6482 0.7040 0.6482 0.8051
No log 8.1667 490 0.8193 0.6599 0.8193 0.9052
No log 8.2 492 0.9313 0.6225 0.9313 0.9650
No log 8.2333 494 0.9139 0.5821 0.9139 0.9560
No log 8.2667 496 0.8163 0.6205 0.8163 0.9035
No log 8.3 498 0.7101 0.6480 0.7101 0.8427
0.4217 8.3333 500 0.6521 0.6598 0.6521 0.8075
0.4217 8.3667 502 0.6360 0.6567 0.6360 0.7975
0.4217 8.4 504 0.6337 0.6833 0.6337 0.7961
0.4217 8.4333 506 0.6801 0.6674 0.6801 0.8247
0.4217 8.4667 508 0.8409 0.6212 0.8409 0.9170
0.4217 8.5 510 0.9761 0.5872 0.9761 0.9880
0.4217 8.5333 512 0.9656 0.5561 0.9656 0.9826
0.4217 8.5667 514 0.8703 0.5491 0.8703 0.9329
0.4217 8.6 516 0.7608 0.5987 0.7608 0.8722

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k12_task1_organization

Finetuned
(4023)
this model