ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k3_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8946
  • Qwk: 0.5882
  • Mse: 0.8946
  • Rmse: 0.9458

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1538 2 6.8428 0.0239 6.8428 2.6159
No log 0.3077 4 4.5021 0.0887 4.5021 2.1218
No log 0.4615 6 4.0366 -0.0505 4.0366 2.0091
No log 0.6154 8 3.0564 0.0488 3.0564 1.7482
No log 0.7692 10 1.9015 0.1197 1.9015 1.3789
No log 0.9231 12 1.6658 0.1852 1.6658 1.2907
No log 1.0769 14 1.6122 0.1818 1.6122 1.2697
No log 1.2308 16 1.8762 0.3158 1.8762 1.3697
No log 1.3846 18 2.0234 0.3194 2.0234 1.4225
No log 1.5385 20 1.7834 0.3529 1.7834 1.3355
No log 1.6923 22 1.6465 0.4091 1.6465 1.2831
No log 1.8462 24 1.5230 0.4091 1.5230 1.2341
No log 2.0 26 1.7321 0.3971 1.7321 1.3161
No log 2.1538 28 1.7991 0.3913 1.7991 1.3413
No log 2.3077 30 1.6252 0.3852 1.6252 1.2748
No log 2.4615 32 1.4850 0.4179 1.4850 1.2186
No log 2.6154 34 1.3477 0.4375 1.3477 1.1609
No log 2.7692 36 1.6961 0.4 1.6961 1.3023
No log 2.9231 38 2.1261 0.3436 2.1261 1.4581
No log 3.0769 40 1.5434 0.4085 1.5434 1.2423
No log 3.2308 42 1.1633 0.4132 1.1633 1.0786
No log 3.3846 44 1.2104 0.4500 1.2104 1.1002
No log 3.5385 46 1.1349 0.5 1.1349 1.0653
No log 3.6923 48 1.2845 0.4848 1.2845 1.1334
No log 3.8462 50 2.5372 0.2286 2.5372 1.5929
No log 4.0 52 3.0658 0.1702 3.0658 1.7509
No log 4.1538 54 2.6705 0.1437 2.6704 1.6342
No log 4.3077 56 1.9037 0.3824 1.9037 1.3797
No log 4.4615 58 1.3691 0.4065 1.3691 1.1701
No log 4.6154 60 1.1800 0.3866 1.1800 1.0863
No log 4.7692 62 1.1167 0.4262 1.1167 1.0568
No log 4.9231 64 1.2195 0.4697 1.2195 1.1043
No log 5.0769 66 1.1301 0.4930 1.1301 1.0630
No log 5.2308 68 0.9582 0.6883 0.9582 0.9789
No log 5.3846 70 0.8268 0.7484 0.8268 0.9093
No log 5.5385 72 0.8712 0.7206 0.8712 0.9334
No log 5.6923 74 0.9046 0.6618 0.9046 0.9511
No log 5.8462 76 0.9697 0.6222 0.9697 0.9848
No log 6.0 78 0.8612 0.7153 0.8612 0.9280
No log 6.1538 80 0.8740 0.6528 0.8740 0.9349
No log 6.3077 82 1.0059 0.5874 1.0059 1.0030
No log 6.4615 84 0.9768 0.5468 0.9768 0.9883
No log 6.6154 86 0.9341 0.6412 0.9341 0.9665
No log 6.7692 88 0.9775 0.6406 0.9775 0.9887
No log 6.9231 90 0.9260 0.6154 0.9260 0.9623
No log 7.0769 92 0.9807 0.5441 0.9807 0.9903
No log 7.2308 94 1.2336 0.5638 1.2336 1.1107
No log 7.3846 96 1.0473 0.6164 1.0473 1.0234
No log 7.5385 98 0.7946 0.6906 0.7946 0.8914
No log 7.6923 100 1.0384 0.5547 1.0384 1.0190
No log 7.8462 102 1.1043 0.5441 1.1043 1.0509
No log 8.0 104 0.8678 0.7206 0.8678 0.9316
No log 8.1538 106 0.8740 0.6533 0.8740 0.9349
No log 8.3077 108 1.1798 0.6424 1.1798 1.0862
No log 8.4615 110 1.0784 0.6707 1.0784 1.0385
No log 8.6154 112 0.7838 0.6980 0.7838 0.8853
No log 8.7692 114 0.9499 0.6429 0.9499 0.9746
No log 8.9231 116 1.0250 0.6197 1.0250 1.0124
No log 9.0769 118 0.8178 0.7324 0.8178 0.9043
No log 9.2308 120 0.7157 0.7162 0.7157 0.8460
No log 9.3846 122 0.8154 0.6710 0.8154 0.9030
No log 9.5385 124 0.8174 0.6839 0.8174 0.9041
No log 9.6923 126 0.7268 0.6939 0.7268 0.8525
No log 9.8462 128 0.7653 0.7101 0.7653 0.8748
No log 10.0 130 0.9043 0.6567 0.9043 0.9510
No log 10.1538 132 0.8378 0.6471 0.8378 0.9153
No log 10.3077 134 0.7850 0.6857 0.7850 0.8860
No log 10.4615 136 1.1076 0.5921 1.1076 1.0524
No log 10.6154 138 1.2659 0.5912 1.2659 1.1251
No log 10.7692 140 1.2250 0.5405 1.2250 1.1068
No log 10.9231 142 1.0501 0.5734 1.0501 1.0247
No log 11.0769 144 0.9189 0.5693 0.9189 0.9586
No log 11.2308 146 0.8663 0.6269 0.8663 0.9307
No log 11.3846 148 0.8359 0.6519 0.8359 0.9143
No log 11.5385 150 0.8045 0.6528 0.8045 0.8969
No log 11.6923 152 0.8395 0.7117 0.8395 0.9163
No log 11.8462 154 0.8072 0.7117 0.8072 0.8984
No log 12.0 156 0.8027 0.7590 0.8027 0.8959
No log 12.1538 158 0.7049 0.7711 0.7049 0.8396
No log 12.3077 160 0.6984 0.7297 0.6984 0.8357
No log 12.4615 162 0.7306 0.6853 0.7306 0.8548
No log 12.6154 164 0.7792 0.6974 0.7792 0.8827
No log 12.7692 166 0.7513 0.6993 0.7513 0.8668
No log 12.9231 168 0.7223 0.7297 0.7223 0.8499
No log 13.0769 170 0.7035 0.7532 0.7035 0.8387
No log 13.2308 172 0.7077 0.7413 0.7077 0.8412
No log 13.3846 174 0.7408 0.7246 0.7408 0.8607
No log 13.5385 176 0.7641 0.7299 0.7641 0.8741
No log 13.6923 178 0.8531 0.6569 0.8531 0.9236
No log 13.8462 180 0.8211 0.6715 0.8211 0.9061
No log 14.0 182 0.7079 0.7550 0.7079 0.8413
No log 14.1538 184 0.7240 0.7320 0.7240 0.8509
No log 14.3077 186 0.7305 0.7682 0.7305 0.8547
No log 14.4615 188 0.7924 0.7391 0.7924 0.8902
No log 14.6154 190 0.7850 0.7465 0.7850 0.8860
No log 14.7692 192 0.7824 0.7338 0.7824 0.8845
No log 14.9231 194 0.8029 0.7429 0.8029 0.8960
No log 15.0769 196 0.9065 0.6370 0.9065 0.9521
No log 15.2308 198 0.8729 0.6767 0.8729 0.9343
No log 15.3846 200 0.8091 0.7153 0.8091 0.8995
No log 15.5385 202 0.8516 0.6423 0.8516 0.9228
No log 15.6923 204 0.9881 0.5248 0.9881 0.9940
No log 15.8462 206 0.9133 0.5429 0.9133 0.9556
No log 16.0 208 0.7900 0.7153 0.7900 0.8888
No log 16.1538 210 0.9596 0.6074 0.9596 0.9796
No log 16.3077 212 1.0600 0.6074 1.0600 1.0296
No log 16.4615 214 0.9642 0.6515 0.9642 0.9820
No log 16.6154 216 0.9197 0.5692 0.9197 0.9590
No log 16.7692 218 0.9184 0.5652 0.9184 0.9583
No log 16.9231 220 0.8202 0.6892 0.8202 0.9056
No log 17.0769 222 0.7842 0.7152 0.7842 0.8855
No log 17.2308 224 0.7879 0.7027 0.7879 0.8877
No log 17.3846 226 0.8063 0.6901 0.8063 0.8979
No log 17.5385 228 0.8717 0.5588 0.8717 0.9336
No log 17.6923 230 0.9102 0.5522 0.9102 0.9541
No log 17.8462 232 0.8807 0.5630 0.8807 0.9385
No log 18.0 234 0.8189 0.6370 0.8189 0.9049
No log 18.1538 236 0.7836 0.7234 0.7836 0.8852
No log 18.3077 238 0.7667 0.7059 0.7667 0.8756
No log 18.4615 240 0.7957 0.6928 0.7957 0.8920
No log 18.6154 242 0.8434 0.6309 0.8434 0.9184
No log 18.7692 244 0.7726 0.6974 0.7726 0.8790
No log 18.9231 246 0.7336 0.72 0.7336 0.8565
No log 19.0769 248 0.7155 0.7417 0.7155 0.8459
No log 19.2308 250 0.7368 0.7417 0.7368 0.8584
No log 19.3846 252 0.7777 0.6906 0.7777 0.8819
No log 19.5385 254 0.8032 0.6667 0.8032 0.8962
No log 19.6923 256 0.8333 0.6963 0.8333 0.9128
No log 19.8462 258 0.8480 0.6815 0.8480 0.9209
No log 20.0 260 0.8287 0.6815 0.8287 0.9104
No log 20.1538 262 0.7740 0.7050 0.7740 0.8798
No log 20.3077 264 0.7285 0.7297 0.7285 0.8535
No log 20.4615 266 0.7356 0.72 0.7356 0.8577
No log 20.6154 268 0.7556 0.7114 0.7556 0.8692
No log 20.7692 270 0.7455 0.6986 0.7455 0.8634
No log 20.9231 272 0.7499 0.6620 0.7499 0.8660
No log 21.0769 274 0.7837 0.6377 0.7837 0.8853
No log 21.2308 276 0.8131 0.6165 0.8131 0.9017
No log 21.3846 278 0.8477 0.6087 0.8477 0.9207
No log 21.5385 280 0.9149 0.5693 0.9149 0.9565
No log 21.6923 282 0.8538 0.6241 0.8538 0.9240
No log 21.8462 284 0.7814 0.6944 0.7814 0.8840
No log 22.0 286 0.7635 0.7133 0.7635 0.8738
No log 22.1538 288 0.7563 0.7162 0.7563 0.8697
No log 22.3077 290 0.7847 0.7114 0.7847 0.8859
No log 22.4615 292 0.8298 0.6803 0.8298 0.9109
No log 22.6154 294 0.8460 0.6712 0.8460 0.9198
No log 22.7692 296 0.8437 0.6087 0.8437 0.9185
No log 22.9231 298 0.8024 0.6324 0.8024 0.8958
No log 23.0769 300 0.7702 0.6569 0.7702 0.8776
No log 23.2308 302 0.7798 0.6667 0.7798 0.8830
No log 23.3846 304 0.7906 0.7123 0.7906 0.8892
No log 23.5385 306 0.7945 0.7114 0.7945 0.8914
No log 23.6923 308 0.7708 0.7320 0.7708 0.8780
No log 23.8462 310 0.8147 0.6993 0.8147 0.9026
No log 24.0 312 0.8476 0.6892 0.8476 0.9206
No log 24.1538 314 0.8476 0.7034 0.8476 0.9206
No log 24.3077 316 0.8369 0.7285 0.8369 0.9148
No log 24.4615 318 0.8501 0.6575 0.8501 0.9220
No log 24.6154 320 0.8553 0.6621 0.8553 0.9248
No log 24.7692 322 0.8505 0.6815 0.8505 0.9222
No log 24.9231 324 0.8411 0.7111 0.8411 0.9171
No log 25.0769 326 0.8240 0.6667 0.8240 0.9077
No log 25.2308 328 0.7548 0.6857 0.7548 0.8688
No log 25.3846 330 0.7506 0.7320 0.7506 0.8664
No log 25.5385 332 0.7388 0.7320 0.7388 0.8596
No log 25.6923 334 0.7016 0.7564 0.7016 0.8376
No log 25.8462 336 0.7161 0.7532 0.7161 0.8462
No log 26.0 338 0.7316 0.7403 0.7316 0.8553
No log 26.1538 340 0.7422 0.7273 0.7422 0.8615
No log 26.3077 342 0.7771 0.7532 0.7771 0.8815
No log 26.4615 344 0.7958 0.7451 0.7958 0.8921
No log 26.6154 346 0.7906 0.7075 0.7906 0.8891
No log 26.7692 348 0.7909 0.6619 0.7909 0.8893
No log 26.9231 350 0.8014 0.6519 0.8014 0.8952
No log 27.0769 352 0.7955 0.6912 0.7955 0.8919
No log 27.2308 354 0.7707 0.6763 0.7707 0.8779
No log 27.3846 356 0.7449 0.7027 0.7449 0.8631
No log 27.5385 358 0.7307 0.7417 0.7307 0.8548
No log 27.6923 360 0.7475 0.72 0.7475 0.8646
No log 27.8462 362 0.7560 0.7297 0.7560 0.8695
No log 28.0 364 0.7889 0.6815 0.7889 0.8882
No log 28.1538 366 0.8482 0.6618 0.8482 0.9210
No log 28.3077 368 0.8500 0.6617 0.8500 0.9220
No log 28.4615 370 0.8409 0.6715 0.8409 0.9170
No log 28.6154 372 0.8745 0.6212 0.8745 0.9351
No log 28.7692 374 0.8818 0.6418 0.8818 0.9391
No log 28.9231 376 0.8874 0.6466 0.8874 0.9420
No log 29.0769 378 0.9130 0.6418 0.9130 0.9555
No log 29.2308 380 0.9543 0.6418 0.9543 0.9769
No log 29.3846 382 0.9395 0.6260 0.9395 0.9693
No log 29.5385 384 0.9142 0.5512 0.9142 0.9561
No log 29.6923 386 0.8939 0.5692 0.8939 0.9455
No log 29.8462 388 0.8641 0.6143 0.8641 0.9296
No log 30.0 390 0.8137 0.7114 0.8137 0.9021
No log 30.1538 392 0.7482 0.7467 0.7482 0.8650
No log 30.3077 394 0.7252 0.7285 0.7252 0.8516
No log 30.4615 396 0.7099 0.7613 0.7099 0.8425
No log 30.6154 398 0.7063 0.7484 0.7063 0.8404
No log 30.7692 400 0.7139 0.7484 0.7139 0.8449
No log 30.9231 402 0.7330 0.7067 0.7330 0.8561
No log 31.0769 404 0.7677 0.7550 0.7677 0.8762
No log 31.2308 406 0.8487 0.6338 0.8487 0.9213
No log 31.3846 408 0.8879 0.6099 0.8879 0.9423
No log 31.5385 410 0.8900 0.5985 0.8900 0.9434
No log 31.6923 412 0.8490 0.6765 0.8490 0.9214
No log 31.8462 414 0.8254 0.6917 0.8254 0.9085
No log 32.0 416 0.8024 0.7101 0.8024 0.8958
No log 32.1538 418 0.8033 0.6986 0.8033 0.8963
No log 32.3077 420 0.8425 0.7123 0.8425 0.9179
No log 32.4615 422 0.8598 0.6944 0.8598 0.9273
No log 32.6154 424 0.8684 0.6765 0.8684 0.9319
No log 32.7692 426 0.8978 0.6567 0.8978 0.9475
No log 32.9231 428 0.8984 0.6176 0.8984 0.9478
No log 33.0769 430 0.8284 0.6815 0.8284 0.9101
No log 33.2308 432 0.7750 0.7211 0.7750 0.8803
No log 33.3846 434 0.8151 0.6712 0.8151 0.9028
No log 33.5385 436 0.8453 0.6389 0.8453 0.9194
No log 33.6923 438 0.7958 0.6892 0.7958 0.8921
No log 33.8462 440 0.7471 0.7517 0.7471 0.8644
No log 34.0 442 0.7803 0.6761 0.7803 0.8833
No log 34.1538 444 0.7827 0.6761 0.7827 0.8847
No log 34.3077 446 0.7385 0.7260 0.7385 0.8594
No log 34.4615 448 0.7136 0.7248 0.7136 0.8448
No log 34.6154 450 0.7258 0.7248 0.7258 0.8519
No log 34.7692 452 0.7641 0.7333 0.7641 0.8741
No log 34.9231 454 0.7972 0.7114 0.7972 0.8929
No log 35.0769 456 0.8054 0.7114 0.8054 0.8974
No log 35.2308 458 0.8194 0.7162 0.8194 0.9052
No log 35.3846 460 0.8078 0.7034 0.8078 0.8988
No log 35.5385 462 0.8127 0.7183 0.8127 0.9015
No log 35.6923 464 0.8287 0.7092 0.8287 0.9103
No log 35.8462 466 0.8128 0.7133 0.8128 0.9016
No log 36.0 468 0.7934 0.7432 0.7934 0.8907
No log 36.1538 470 0.7982 0.7432 0.7982 0.8934
No log 36.3077 472 0.8095 0.7432 0.8095 0.8997
No log 36.4615 474 0.8222 0.7075 0.8222 0.9068
No log 36.6154 476 0.8240 0.6993 0.8240 0.9077
No log 36.7692 478 0.8105 0.6812 0.8105 0.9003
No log 36.9231 480 0.8071 0.7206 0.8071 0.8984
No log 37.0769 482 0.8359 0.6866 0.8359 0.9143
No log 37.2308 484 0.8364 0.6917 0.8364 0.9145
No log 37.3846 486 0.8166 0.6912 0.8166 0.9037
No log 37.5385 488 0.8143 0.6812 0.8143 0.9024
No log 37.6923 490 0.7997 0.7027 0.7997 0.8943
No log 37.8462 492 0.7706 0.7333 0.7706 0.8779
No log 38.0 494 0.7682 0.7484 0.7682 0.8765
No log 38.1538 496 0.7842 0.7308 0.7842 0.8856
No log 38.3077 498 0.7789 0.7190 0.7789 0.8826
0.3242 38.4615 500 0.7697 0.7712 0.7697 0.8773
0.3242 38.6154 502 0.7827 0.7183 0.7827 0.8847
0.3242 38.7692 504 0.8030 0.6957 0.8030 0.8961
0.3242 38.9231 506 0.8138 0.6715 0.8138 0.9021
0.3242 39.0769 508 0.8376 0.6423 0.8376 0.9152
0.3242 39.2308 510 0.8946 0.5882 0.8946 0.9458

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k3_task1_organization

Finetuned
(4023)
this model