ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k6_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6094
  • Qwk: 0.7586
  • Mse: 0.6094
  • Rmse: 0.7806

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0714 2 6.7446 0.0308 6.7446 2.5970
No log 0.1429 4 4.4270 0.0342 4.4270 2.1040
No log 0.2143 6 3.1924 0.0 3.1924 1.7867
No log 0.2857 8 2.2548 0.1507 2.2548 1.5016
No log 0.3571 10 2.4802 0.1429 2.4802 1.5749
No log 0.4286 12 2.1155 0.1406 2.1155 1.4545
No log 0.5 14 1.9806 0.2114 1.9806 1.4073
No log 0.5714 16 2.3859 0.1333 2.3859 1.5446
No log 0.6429 18 2.5757 0.1366 2.5757 1.6049
No log 0.7143 20 2.8047 0.1310 2.8047 1.6747
No log 0.7857 22 2.3858 0.1500 2.3858 1.5446
No log 0.8571 24 1.7168 0.4252 1.7168 1.3103
No log 0.9286 26 1.4714 0.2000 1.4714 1.2130
No log 1.0 28 1.5468 0.1321 1.5468 1.2437
No log 1.0714 30 1.5943 0.2018 1.5943 1.2627
No log 1.1429 32 1.6585 0.3248 1.6585 1.2878
No log 1.2143 34 1.7289 0.4127 1.7289 1.3149
No log 1.2857 36 2.1963 0.2000 2.1963 1.4820
No log 1.3571 38 2.8634 0.1503 2.8634 1.6922
No log 1.4286 40 2.8533 0.1453 2.8533 1.6892
No log 1.5 42 2.2977 0.2561 2.2977 1.5158
No log 1.5714 44 2.1259 0.2893 2.1259 1.4580
No log 1.6429 46 1.8704 0.3896 1.8704 1.3676
No log 1.7143 48 1.7639 0.4625 1.7639 1.3281
No log 1.7857 50 1.6704 0.5063 1.6704 1.2924
No log 1.8571 52 1.9191 0.4114 1.9191 1.3853
No log 1.9286 54 2.4653 0.2653 2.4653 1.5701
No log 2.0 56 2.2724 0.3240 2.2724 1.5074
No log 2.0714 58 1.6760 0.4503 1.6760 1.2946
No log 2.1429 60 1.4581 0.4552 1.4581 1.2075
No log 2.2143 62 1.2907 0.5634 1.2907 1.1361
No log 2.2857 64 1.3420 0.5442 1.3420 1.1585
No log 2.3571 66 1.7742 0.4250 1.7742 1.3320
No log 2.4286 68 1.7960 0.3975 1.7960 1.3401
No log 2.5 70 1.3224 0.5658 1.3224 1.1499
No log 2.5714 72 0.8598 0.7075 0.8598 0.9272
No log 2.6429 74 0.9539 0.6525 0.9539 0.9767
No log 2.7143 76 1.0282 0.6087 1.0282 1.0140
No log 2.7857 78 0.8218 0.7190 0.8218 0.9065
No log 2.8571 80 1.0962 0.6353 1.0962 1.0470
No log 2.9286 82 1.6450 0.4973 1.6450 1.2826
No log 3.0 84 1.6676 0.4945 1.6676 1.2914
No log 3.0714 86 1.2070 0.5732 1.2070 1.0987
No log 3.1429 88 0.8619 0.6797 0.8619 0.9284
No log 3.2143 90 0.7326 0.7211 0.7326 0.8559
No log 3.2857 92 0.7543 0.7083 0.7543 0.8685
No log 3.3571 94 0.9142 0.6483 0.9142 0.9562
No log 3.4286 96 1.3131 0.5786 1.3131 1.1459
No log 3.5 98 1.9496 0.4688 1.9496 1.3963
No log 3.5714 100 2.3369 0.4171 2.3369 1.5287
No log 3.6429 102 2.1091 0.4948 2.1091 1.4523
No log 3.7143 104 1.5819 0.5444 1.5819 1.2577
No log 3.7857 106 1.1944 0.6374 1.1944 1.0929
No log 3.8571 108 0.9030 0.7486 0.9030 0.9503
No log 3.9286 110 0.8493 0.75 0.8493 0.9216
No log 4.0 112 0.8443 0.7556 0.8443 0.9189
No log 4.0714 114 0.7772 0.7135 0.7772 0.8816
No log 4.1429 116 0.7791 0.7294 0.7791 0.8827
No log 4.2143 118 0.7731 0.7317 0.7731 0.8792
No log 4.2857 120 0.7223 0.7013 0.7223 0.8499
No log 4.3571 122 0.7249 0.7237 0.7249 0.8514
No log 4.4286 124 0.6982 0.7550 0.6982 0.8356
No log 4.5 126 0.7636 0.7347 0.7636 0.8738
No log 4.5714 128 1.0526 0.5946 1.0526 1.0259
No log 4.6429 130 1.1443 0.6104 1.1443 1.0697
No log 4.7143 132 0.9048 0.6389 0.9048 0.9512
No log 4.7857 134 0.7179 0.7347 0.7179 0.8473
No log 4.8571 136 0.7932 0.7123 0.7932 0.8906
No log 4.9286 138 0.7788 0.7347 0.7788 0.8825
No log 5.0 140 0.9143 0.6867 0.9143 0.9562
No log 5.0714 142 1.6912 0.5 1.6912 1.3005
No log 5.1429 144 2.1645 0.3942 2.1645 1.4712
No log 5.2143 146 1.9030 0.4718 1.9030 1.3795
No log 5.2857 148 1.3170 0.6012 1.3170 1.1476
No log 5.3571 150 0.9139 0.6531 0.9139 0.9560
No log 5.4286 152 1.0294 0.6621 1.0294 1.0146
No log 5.5 154 1.0937 0.6434 1.0937 1.0458
No log 5.5714 156 0.8731 0.7034 0.8731 0.9344
No log 5.6429 158 0.8047 0.6667 0.8047 0.8971
No log 5.7143 160 0.9784 0.6479 0.9784 0.9892
No log 5.7857 162 1.6214 0.5 1.6214 1.2734
No log 5.8571 164 2.0689 0.3784 2.0689 1.4384
No log 5.9286 166 1.8903 0.4396 1.8903 1.3749
No log 6.0 168 1.3028 0.6243 1.3028 1.1414
No log 6.0714 170 0.8324 0.6624 0.8324 0.9124
No log 6.1429 172 0.7542 0.7089 0.7542 0.8685
No log 6.2143 174 0.7321 0.7037 0.7321 0.8556
No log 6.2857 176 0.6978 0.7195 0.6978 0.8353
No log 6.3571 178 0.6995 0.7195 0.6995 0.8364
No log 6.4286 180 0.7096 0.7362 0.7096 0.8424
No log 6.5 182 0.7298 0.7561 0.7298 0.8543
No log 6.5714 184 0.7475 0.7349 0.7475 0.8646
No log 6.6429 186 0.7349 0.7547 0.7349 0.8572
No log 6.7143 188 0.7386 0.7843 0.7386 0.8594
No log 6.7857 190 0.7302 0.7651 0.7302 0.8545
No log 6.8571 192 0.8123 0.7517 0.8123 0.9013
No log 6.9286 194 0.8435 0.7162 0.8435 0.9184
No log 7.0 196 0.7991 0.6914 0.7991 0.8939
No log 7.0714 198 0.9013 0.7006 0.9013 0.9494
No log 7.1429 200 1.1255 0.6519 1.1255 1.0609
No log 7.2143 202 1.2744 0.6129 1.2744 1.1289
No log 7.2857 204 1.0308 0.6629 1.0308 1.0153
No log 7.3571 206 0.8654 0.7059 0.8654 0.9303
No log 7.4286 208 0.6974 0.7582 0.6974 0.8351
No log 7.5 210 0.6465 0.7432 0.6465 0.8040
No log 7.5714 212 0.6594 0.7467 0.6594 0.8121
No log 7.6429 214 0.6675 0.7582 0.6675 0.8170
No log 7.7143 216 0.7734 0.7355 0.7734 0.8794
No log 7.7857 218 0.9907 0.6467 0.9907 0.9953
No log 7.8571 220 1.2477 0.6163 1.2477 1.1170
No log 7.9286 222 1.2088 0.6087 1.2088 1.0995
No log 8.0 224 0.9696 0.6405 0.9696 0.9847
No log 8.0714 226 0.7443 0.7042 0.7443 0.8627
No log 8.1429 228 0.6657 0.7483 0.6657 0.8159
No log 8.2143 230 0.6923 0.775 0.6923 0.8321
No log 8.2857 232 0.8390 0.7241 0.8390 0.9160
No log 8.3571 234 1.2097 0.6486 1.2097 1.0999
No log 8.4286 236 1.3728 0.6186 1.3728 1.1717
No log 8.5 238 1.1867 0.6802 1.1867 1.0893
No log 8.5714 240 0.8762 0.7391 0.8762 0.9361
No log 8.6429 242 0.6573 0.7719 0.6573 0.8107
No log 8.7143 244 0.6589 0.7582 0.6589 0.8117
No log 8.7857 246 0.7722 0.7183 0.7722 0.8788
No log 8.8571 248 0.9594 0.6622 0.9594 0.9795
No log 8.9286 250 1.1813 0.6040 1.1813 1.0869
No log 9.0 252 1.1389 0.6040 1.1389 1.0672
No log 9.0714 254 0.9411 0.6351 0.9411 0.9701
No log 9.1429 256 0.6940 0.7310 0.6940 0.8330
No log 9.2143 258 0.6033 0.7733 0.6033 0.7767
No log 9.2857 260 0.5836 0.8199 0.5836 0.7639
No log 9.3571 262 0.5982 0.7765 0.5982 0.7735
No log 9.4286 264 0.6288 0.7657 0.6288 0.7930
No log 9.5 266 0.6812 0.7797 0.6812 0.8254
No log 9.5714 268 0.6846 0.7614 0.6846 0.8274
No log 9.6429 270 0.6720 0.7931 0.6720 0.8197
No log 9.7143 272 0.6494 0.7931 0.6494 0.8059
No log 9.7857 274 0.6735 0.7746 0.6735 0.8207
No log 9.8571 276 0.6407 0.8046 0.6407 0.8004
No log 9.9286 278 0.6450 0.7895 0.6450 0.8031
No log 10.0 280 0.6803 0.7785 0.6803 0.8248
No log 10.0714 282 0.7038 0.7792 0.7038 0.8389
No log 10.1429 284 0.7399 0.7545 0.7399 0.8602
No log 10.2143 286 0.8014 0.7356 0.8014 0.8952
No log 10.2857 288 0.8809 0.7543 0.8809 0.9386
No log 10.3571 290 0.8884 0.7416 0.8884 0.9426
No log 10.4286 292 0.7235 0.7399 0.7235 0.8506
No log 10.5 294 0.6753 0.7647 0.6753 0.8218
No log 10.5714 296 0.6515 0.7836 0.6515 0.8072
No log 10.6429 298 0.6849 0.7529 0.6849 0.8276
No log 10.7143 300 0.8386 0.7598 0.8386 0.9157
No log 10.7857 302 1.0459 0.6845 1.0459 1.0227
No log 10.8571 304 0.9526 0.7263 0.9526 0.9760
No log 10.9286 306 0.7744 0.7582 0.7744 0.8800
No log 11.0 308 0.6644 0.8023 0.6644 0.8151
No log 11.0714 310 0.6444 0.8090 0.6444 0.8027
No log 11.1429 312 0.6097 0.8208 0.6097 0.7808
No log 11.2143 314 0.5966 0.7929 0.5966 0.7724
No log 11.2857 316 0.5846 0.8024 0.5846 0.7646
No log 11.3571 318 0.6064 0.7778 0.6064 0.7787
No log 11.4286 320 0.6521 0.7799 0.6521 0.8075
No log 11.5 322 0.6562 0.7711 0.6562 0.8100
No log 11.5714 324 0.7117 0.7630 0.7117 0.8436
No log 11.6429 326 0.7110 0.7771 0.7110 0.8432
No log 11.7143 328 0.6431 0.7531 0.6431 0.8020
No log 11.7857 330 0.6555 0.7815 0.6555 0.8096
No log 11.8571 332 0.7140 0.7742 0.7140 0.8450
No log 11.9286 334 0.7446 0.7595 0.7446 0.8629
No log 12.0 336 0.8328 0.6994 0.8328 0.9126
No log 12.0714 338 0.9225 0.6982 0.9225 0.9605
No log 12.1429 340 0.9491 0.6927 0.9491 0.9742
No log 12.2143 342 0.8924 0.7111 0.8924 0.9447
No log 12.2857 344 0.7617 0.7701 0.7617 0.8728
No log 12.3571 346 0.6334 0.7673 0.6334 0.7959
No log 12.4286 348 0.6174 0.8 0.6174 0.7857
No log 12.5 350 0.6261 0.7568 0.6261 0.7913
No log 12.5714 352 0.6139 0.7785 0.6139 0.7835
No log 12.6429 354 0.6627 0.7582 0.6627 0.8140
No log 12.7143 356 0.8770 0.6867 0.8770 0.9365
No log 12.7857 358 1.0358 0.6704 1.0358 1.0177
No log 12.8571 360 0.9545 0.6704 0.9545 0.9770
No log 12.9286 362 0.7313 0.7624 0.7313 0.8552
No log 13.0 364 0.5915 0.7657 0.5915 0.7691
No log 13.0714 366 0.5777 0.7882 0.5777 0.7600
No log 13.1429 368 0.5950 0.7882 0.5950 0.7714
No log 13.2143 370 0.6555 0.8 0.6555 0.8096
No log 13.2857 372 0.8171 0.7582 0.8171 0.9039
No log 13.3571 374 0.9144 0.7065 0.9144 0.9562
No log 13.4286 376 1.0149 0.6885 1.0149 1.0074
No log 13.5 378 0.9628 0.7213 0.9628 0.9812
No log 13.5714 380 0.7700 0.7514 0.7700 0.8775
No log 13.6429 382 0.6581 0.7578 0.6581 0.8112
No log 13.7143 384 0.6044 0.7895 0.6044 0.7774
No log 13.7857 386 0.5963 0.7949 0.5963 0.7722
No log 13.8571 388 0.6447 0.7811 0.6447 0.8029
No log 13.9286 390 0.6793 0.7811 0.6793 0.8242
No log 14.0 392 0.7859 0.7239 0.7859 0.8865
No log 14.0714 394 0.8962 0.7117 0.8962 0.9467
No log 14.1429 396 0.9004 0.7117 0.9004 0.9489
No log 14.2143 398 0.7952 0.7152 0.7952 0.8917
No log 14.2857 400 0.7059 0.7636 0.7059 0.8401
No log 14.3571 402 0.7341 0.7561 0.7341 0.8568
No log 14.4286 404 0.7392 0.7561 0.7392 0.8598
No log 14.5 406 0.7080 0.775 0.7080 0.8414
No log 14.5714 408 0.6974 0.7811 0.6974 0.8351
No log 14.6429 410 0.7106 0.7765 0.7106 0.8430
No log 14.7143 412 0.7108 0.7816 0.7108 0.8431
No log 14.7857 414 0.7083 0.7865 0.7083 0.8416
No log 14.8571 416 0.6633 0.7811 0.6633 0.8144
No log 14.9286 418 0.6748 0.7683 0.6748 0.8215
No log 15.0 420 0.6889 0.7730 0.6889 0.8300
No log 15.0714 422 0.6716 0.7875 0.6716 0.8195
No log 15.1429 424 0.6671 0.7922 0.6671 0.8168
No log 15.2143 426 0.7080 0.7871 0.7080 0.8414
No log 15.2857 428 0.8277 0.7320 0.8277 0.9098
No log 15.3571 430 0.9219 0.6335 0.9219 0.9602
No log 15.4286 432 0.8931 0.6503 0.8931 0.9450
No log 15.5 434 0.8007 0.6918 0.8007 0.8948
No log 15.5714 436 0.7014 0.7792 0.7014 0.8375
No log 15.6429 438 0.6554 0.7922 0.6554 0.8095
No log 15.7143 440 0.6221 0.8025 0.6221 0.7887
No log 15.7857 442 0.5852 0.8125 0.5852 0.7650
No log 15.8571 444 0.5825 0.8 0.5825 0.7632
No log 15.9286 446 0.5905 0.7976 0.5905 0.7684
No log 16.0 448 0.5930 0.7976 0.5930 0.7701
No log 16.0714 450 0.6122 0.8101 0.6122 0.7824
No log 16.1429 452 0.6297 0.7867 0.6297 0.7936
No log 16.2143 454 0.6824 0.7815 0.6824 0.8261
No log 16.2857 456 0.7784 0.7451 0.7784 0.8823
No log 16.3571 458 0.8238 0.7673 0.8238 0.9076
No log 16.4286 460 0.8046 0.7349 0.8046 0.8970
No log 16.5 462 0.7308 0.7514 0.7308 0.8549
No log 16.5714 464 0.6410 0.7836 0.6410 0.8006
No log 16.6429 466 0.5980 0.7975 0.5980 0.7733
No log 16.7143 468 0.6143 0.8000 0.6143 0.7838
No log 16.7857 470 0.6771 0.7662 0.6771 0.8228
No log 16.8571 472 0.7063 0.7550 0.7063 0.8404
No log 16.9286 474 0.7503 0.75 0.7503 0.8662
No log 17.0 476 0.7370 0.7582 0.7370 0.8585
No log 17.0714 478 0.6800 0.7922 0.6800 0.8246
No log 17.1429 480 0.6218 0.7922 0.6218 0.7885
No log 17.2143 482 0.6197 0.7703 0.6197 0.7872
No log 17.2857 484 0.6439 0.7397 0.6439 0.8025
No log 17.3571 486 0.6464 0.7397 0.6464 0.8040
No log 17.4286 488 0.6541 0.7397 0.6541 0.8087
No log 17.5 490 0.6596 0.7586 0.6596 0.8122
No log 17.5714 492 0.6627 0.7448 0.6627 0.8141
No log 17.6429 494 0.7043 0.7922 0.7043 0.8392
No log 17.7143 496 0.7556 0.8025 0.7556 0.8692
No log 17.7857 498 0.7376 0.8025 0.7376 0.8588
0.3917 17.8571 500 0.6682 0.7949 0.6682 0.8174
0.3917 17.9286 502 0.6241 0.7619 0.6241 0.7900
0.3917 18.0 504 0.6042 0.7671 0.6042 0.7773
0.3917 18.0714 506 0.5938 0.7586 0.5938 0.7706
0.3917 18.1429 508 0.6040 0.7586 0.6040 0.7772
0.3917 18.2143 510 0.6094 0.7586 0.6094 0.7806

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k6_task1_organization

Finetuned
(4019)
this model