ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k15_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7086
  • Qwk: 0.4473
  • Mse: 0.7086
  • Rmse: 0.8418

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0417 2 4.2566 0.0087 4.2566 2.0631
No log 0.0833 4 2.5674 0.0318 2.5674 1.6023
No log 0.125 6 1.9612 0.0247 1.9612 1.4004
No log 0.1667 8 1.6540 0.1060 1.6540 1.2861
No log 0.2083 10 1.1110 0.2947 1.1110 1.0541
No log 0.25 12 1.0199 0.2441 1.0199 1.0099
No log 0.2917 14 1.1061 0.2265 1.1061 1.0517
No log 0.3333 16 1.1534 0.1685 1.1534 1.0740
No log 0.375 18 1.1824 0.2030 1.1824 1.0874
No log 0.4167 20 1.3960 0.0513 1.3960 1.1815
No log 0.4583 22 1.3540 0.0513 1.3540 1.1636
No log 0.5 24 1.1962 0.2030 1.1962 1.0937
No log 0.5417 26 1.0949 0.2120 1.0949 1.0464
No log 0.5833 28 1.0380 0.2192 1.0380 1.0188
No log 0.625 30 1.0352 0.2192 1.0352 1.0175
No log 0.6667 32 1.1354 0.2100 1.1354 1.0655
No log 0.7083 34 1.6188 0.0530 1.6188 1.2723
No log 0.75 36 1.9847 0.0916 1.9847 1.4088
No log 0.7917 38 1.8833 0.0733 1.8833 1.3723
No log 0.8333 40 1.4662 0.1084 1.4662 1.2109
No log 0.875 42 1.1759 0.1738 1.1759 1.0844
No log 0.9167 44 1.0494 0.1794 1.0494 1.0244
No log 0.9583 46 1.0343 0.2265 1.0343 1.0170
No log 1.0 48 1.0436 0.2316 1.0436 1.0215
No log 1.0417 50 1.1155 0.1779 1.1155 1.0562
No log 1.0833 52 1.2898 0.1817 1.2898 1.1357
No log 1.125 54 1.4072 0.1060 1.4072 1.1862
No log 1.1667 56 1.4788 0.0850 1.4788 1.2160
No log 1.2083 58 1.3488 0.1262 1.3488 1.1614
No log 1.25 60 1.2124 0.2225 1.2124 1.1011
No log 1.2917 62 1.2823 0.2526 1.2823 1.1324
No log 1.3333 64 1.4105 0.2632 1.4105 1.1876
No log 1.375 66 1.4834 0.1822 1.4834 1.2179
No log 1.4167 68 1.1962 0.2583 1.1962 1.0937
No log 1.4583 70 1.0957 0.3702 1.0957 1.0468
No log 1.5 72 1.0321 0.3352 1.0321 1.0159
No log 1.5417 74 1.1563 0.2899 1.1563 1.0753
No log 1.5833 76 1.1133 0.3300 1.1133 1.0551
No log 1.625 78 1.0084 0.2892 1.0084 1.0042
No log 1.6667 80 0.9763 0.2229 0.9763 0.9881
No log 1.7083 82 0.9717 0.2326 0.9717 0.9858
No log 1.75 84 0.9568 0.2479 0.9568 0.9781
No log 1.7917 86 0.9528 0.3004 0.9528 0.9761
No log 1.8333 88 0.9692 0.4237 0.9692 0.9845
No log 1.875 90 1.0759 0.3675 1.0759 1.0373
No log 1.9167 92 1.0808 0.4229 1.0808 1.0396
No log 1.9583 94 0.9284 0.4331 0.9284 0.9635
No log 2.0 96 0.8415 0.4645 0.8415 0.9174
No log 2.0417 98 0.9548 0.3987 0.9548 0.9772
No log 2.0833 100 0.9639 0.4785 0.9639 0.9818
No log 2.125 102 0.8957 0.4935 0.8957 0.9464
No log 2.1667 104 0.8285 0.3559 0.8285 0.9102
No log 2.2083 106 0.8124 0.3856 0.8124 0.9013
No log 2.25 108 0.8581 0.4507 0.8581 0.9263
No log 2.2917 110 0.8545 0.4856 0.8545 0.9244
No log 2.3333 112 0.8046 0.5567 0.8046 0.8970
No log 2.375 114 0.8307 0.5178 0.8307 0.9114
No log 2.4167 116 0.8154 0.5372 0.8154 0.9030
No log 2.4583 118 0.8581 0.5977 0.8581 0.9263
No log 2.5 120 0.9124 0.4731 0.9124 0.9552
No log 2.5417 122 0.8455 0.5458 0.8455 0.9195
No log 2.5833 124 0.7847 0.5513 0.7847 0.8859
No log 2.625 126 0.7710 0.5407 0.7710 0.8781
No log 2.6667 128 0.7540 0.5871 0.7540 0.8683
No log 2.7083 130 0.7444 0.5881 0.7444 0.8628
No log 2.75 132 0.7322 0.6157 0.7322 0.8557
No log 2.7917 134 0.8121 0.5963 0.8121 0.9012
No log 2.8333 136 0.8567 0.4465 0.8567 0.9256
No log 2.875 138 0.8291 0.4918 0.8291 0.9105
No log 2.9167 140 0.8044 0.4870 0.8044 0.8969
No log 2.9583 142 0.7400 0.5085 0.7400 0.8602
No log 3.0 144 0.7095 0.5287 0.7095 0.8423
No log 3.0417 146 0.6867 0.5368 0.6867 0.8287
No log 3.0833 148 0.8081 0.5137 0.8081 0.8989
No log 3.125 150 0.9331 0.4668 0.9331 0.9660
No log 3.1667 152 0.8489 0.5137 0.8489 0.9213
No log 3.2083 154 0.7372 0.4712 0.7372 0.8586
No log 3.25 156 0.7076 0.5331 0.7076 0.8412
No log 3.2917 158 0.6939 0.5528 0.6939 0.8330
No log 3.3333 160 0.6896 0.5844 0.6896 0.8304
No log 3.375 162 0.7212 0.5494 0.7212 0.8492
No log 3.4167 164 0.7202 0.5563 0.7202 0.8487
No log 3.4583 166 0.7428 0.5669 0.7428 0.8619
No log 3.5 168 0.7343 0.5877 0.7343 0.8569
No log 3.5417 170 0.7587 0.6045 0.7587 0.8710
No log 3.5833 172 0.7918 0.5466 0.7918 0.8898
No log 3.625 174 0.7829 0.5688 0.7829 0.8848
No log 3.6667 176 0.7378 0.5905 0.7378 0.8589
No log 3.7083 178 0.7350 0.5909 0.7350 0.8573
No log 3.75 180 0.7354 0.5975 0.7354 0.8575
No log 3.7917 182 0.7651 0.5977 0.7651 0.8747
No log 3.8333 184 0.7441 0.6084 0.7441 0.8626
No log 3.875 186 0.7775 0.5183 0.7775 0.8818
No log 3.9167 188 1.0535 0.5242 1.0535 1.0264
No log 3.9583 190 1.1247 0.5422 1.1247 1.0605
No log 4.0 192 0.8851 0.5297 0.8851 0.9408
No log 4.0417 194 0.7978 0.5462 0.7978 0.8932
No log 4.0833 196 0.9621 0.4580 0.9621 0.9808
No log 4.125 198 0.8922 0.4787 0.8922 0.9446
No log 4.1667 200 0.7313 0.5348 0.7313 0.8552
No log 4.2083 202 0.7219 0.5166 0.7219 0.8497
No log 4.25 204 0.7198 0.5163 0.7198 0.8484
No log 4.2917 206 0.7391 0.4729 0.7391 0.8597
No log 4.3333 208 0.8411 0.4460 0.8411 0.9171
No log 4.375 210 0.9082 0.5098 0.9082 0.9530
No log 4.4167 212 0.8403 0.4818 0.8403 0.9167
No log 4.4583 214 0.7570 0.4824 0.7570 0.8701
No log 4.5 216 0.7584 0.5279 0.7584 0.8708
No log 4.5417 218 0.7956 0.5475 0.7956 0.8920
No log 4.5833 220 0.8885 0.4785 0.8885 0.9426
No log 4.625 222 0.8201 0.5358 0.8201 0.9056
No log 4.6667 224 0.7253 0.5422 0.7253 0.8516
No log 4.7083 226 0.6943 0.5063 0.6943 0.8332
No log 4.75 228 0.6801 0.6147 0.6801 0.8247
No log 4.7917 230 0.7758 0.5766 0.7758 0.8808
No log 4.8333 232 0.8014 0.5655 0.8014 0.8952
No log 4.875 234 0.7064 0.6218 0.7064 0.8405
No log 4.9167 236 0.6706 0.6745 0.6706 0.8189
No log 4.9583 238 0.6632 0.5747 0.6632 0.8144
No log 5.0 240 0.6655 0.5735 0.6655 0.8158
No log 5.0417 242 0.6867 0.6133 0.6867 0.8286
No log 5.0833 244 0.7396 0.5254 0.7396 0.8600
No log 5.125 246 0.7314 0.5292 0.7314 0.8552
No log 5.1667 248 0.6902 0.6165 0.6902 0.8308
No log 5.2083 250 0.6694 0.5747 0.6694 0.8181
No log 5.25 252 0.6591 0.6067 0.6591 0.8118
No log 5.2917 254 0.6666 0.6113 0.6666 0.8165
No log 5.3333 256 0.7136 0.5756 0.7136 0.8447
No log 5.375 258 0.7271 0.5745 0.7271 0.8527
No log 5.4167 260 0.6593 0.5966 0.6593 0.8120
No log 5.4583 262 0.6595 0.5536 0.6595 0.8121
No log 5.5 264 0.6459 0.5833 0.6459 0.8037
No log 5.5417 266 0.6430 0.5980 0.6430 0.8019
No log 5.5833 268 0.6399 0.5505 0.6399 0.7999
No log 5.625 270 0.6590 0.5759 0.6590 0.8118
No log 5.6667 272 0.6445 0.5713 0.6445 0.8028
No log 5.7083 274 0.6803 0.5821 0.6803 0.8248
No log 5.75 276 0.7317 0.5877 0.7317 0.8554
No log 5.7917 278 0.6679 0.5895 0.6679 0.8173
No log 5.8333 280 0.6368 0.6032 0.6368 0.7980
No log 5.875 282 0.6325 0.5985 0.6325 0.7953
No log 5.9167 284 0.6510 0.6237 0.6510 0.8068
No log 5.9583 286 0.7029 0.6127 0.7029 0.8384
No log 6.0 288 0.6694 0.6237 0.6694 0.8182
No log 6.0417 290 0.6830 0.5542 0.6830 0.8264
No log 6.0833 292 0.7454 0.5541 0.7454 0.8634
No log 6.125 294 0.6958 0.6144 0.6958 0.8341
No log 6.1667 296 0.6835 0.6129 0.6835 0.8267
No log 6.2083 298 0.6908 0.6129 0.6908 0.8311
No log 6.25 300 0.6819 0.6365 0.6819 0.8258
No log 6.2917 302 0.6988 0.6067 0.6988 0.8360
No log 6.3333 304 0.7065 0.6311 0.7065 0.8406
No log 6.375 306 0.7878 0.5044 0.7878 0.8876
No log 6.4167 308 0.9103 0.5190 0.9103 0.9541
No log 6.4583 310 0.8404 0.5505 0.8404 0.9168
No log 6.5 312 0.7236 0.5774 0.7236 0.8507
No log 6.5417 314 0.7386 0.5828 0.7386 0.8594
No log 6.5833 316 0.7187 0.6206 0.7187 0.8478
No log 6.625 318 0.7151 0.5720 0.7151 0.8457
No log 6.6667 320 0.6991 0.6196 0.6991 0.8361
No log 6.7083 322 0.6948 0.5659 0.6948 0.8336
No log 6.75 324 0.6950 0.5441 0.6950 0.8337
No log 6.7917 326 0.6876 0.4944 0.6876 0.8292
No log 6.8333 328 0.6794 0.4944 0.6794 0.8243
No log 6.875 330 0.6747 0.4944 0.6747 0.8214
No log 6.9167 332 0.6999 0.5581 0.6999 0.8366
No log 6.9583 334 0.6724 0.5581 0.6724 0.8200
No log 7.0 336 0.6377 0.5843 0.6377 0.7986
No log 7.0417 338 0.6431 0.6239 0.6431 0.8019
No log 7.0833 340 0.6469 0.6430 0.6469 0.8043
No log 7.125 342 0.6551 0.6337 0.6551 0.8094
No log 7.1667 344 0.6365 0.6680 0.6365 0.7978
No log 7.2083 346 0.6316 0.6584 0.6316 0.7947
No log 7.25 348 0.6348 0.6006 0.6348 0.7967
No log 7.2917 350 0.6430 0.6602 0.6430 0.8019
No log 7.3333 352 0.6576 0.6415 0.6576 0.8109
No log 7.375 354 0.6865 0.5708 0.6865 0.8286
No log 7.4167 356 0.6770 0.5747 0.6770 0.8228
No log 7.4583 358 0.6939 0.5795 0.6939 0.8330
No log 7.5 360 0.7214 0.5442 0.7214 0.8493
No log 7.5417 362 0.7203 0.5648 0.7203 0.8487
No log 7.5833 364 0.6820 0.6294 0.6820 0.8258
No log 7.625 366 0.6738 0.6774 0.6738 0.8208
No log 7.6667 368 0.6837 0.6685 0.6837 0.8268
No log 7.7083 370 0.6356 0.6699 0.6356 0.7973
No log 7.75 372 0.6774 0.5455 0.6774 0.8230
No log 7.7917 374 0.6869 0.5331 0.6869 0.8288
No log 7.8333 376 0.6620 0.5316 0.6620 0.8136
No log 7.875 378 0.6679 0.5928 0.6679 0.8172
No log 7.9167 380 0.6743 0.6330 0.6743 0.8212
No log 7.9583 382 0.6960 0.6479 0.6960 0.8342
No log 8.0 384 0.6925 0.6266 0.6925 0.8322
No log 8.0417 386 0.7073 0.5872 0.7073 0.8410
No log 8.0833 388 0.6813 0.6235 0.6813 0.8254
No log 8.125 390 0.7034 0.5645 0.7034 0.8387
No log 8.1667 392 0.6999 0.5442 0.6999 0.8366
No log 8.2083 394 0.6573 0.5466 0.6573 0.8107
No log 8.25 396 0.6754 0.6626 0.6754 0.8218
No log 8.2917 398 0.7282 0.6071 0.7282 0.8533
No log 8.3333 400 0.7021 0.6774 0.7021 0.8379
No log 8.375 402 0.6891 0.6724 0.6891 0.8301
No log 8.4167 404 0.7129 0.6266 0.7129 0.8443
No log 8.4583 406 0.7016 0.5983 0.7016 0.8376
No log 8.5 408 0.6616 0.6701 0.6616 0.8134
No log 8.5417 410 0.6504 0.6196 0.6504 0.8065
No log 8.5833 412 0.6634 0.6597 0.6634 0.8145
No log 8.625 414 0.7042 0.5677 0.7042 0.8392
No log 8.6667 416 0.7007 0.5777 0.7007 0.8371
No log 8.7083 418 0.6812 0.5279 0.6812 0.8253
No log 8.75 420 0.6823 0.5708 0.6823 0.8260
No log 8.7917 422 0.6676 0.6167 0.6676 0.8170
No log 8.8333 424 0.6932 0.5498 0.6932 0.8326
No log 8.875 426 0.7063 0.5902 0.7063 0.8404
No log 8.9167 428 0.6844 0.6361 0.6844 0.8273
No log 8.9583 430 0.6652 0.5569 0.6652 0.8156
No log 9.0 432 0.6576 0.5975 0.6576 0.8109
No log 9.0417 434 0.6620 0.6314 0.6620 0.8136
No log 9.0833 436 0.6683 0.6314 0.6683 0.8175
No log 9.125 438 0.6434 0.5949 0.6434 0.8021
No log 9.1667 440 0.6525 0.5188 0.6525 0.8078
No log 9.2083 442 0.6647 0.5558 0.6647 0.8153
No log 9.25 444 0.6887 0.6127 0.6887 0.8299
No log 9.2917 446 0.6524 0.5657 0.6524 0.8077
No log 9.3333 448 0.7017 0.4903 0.7017 0.8377
No log 9.375 450 0.7805 0.5051 0.7805 0.8835
No log 9.4167 452 0.7362 0.5025 0.7362 0.8580
No log 9.4583 454 0.6419 0.5548 0.6419 0.8012
No log 9.5 456 0.6853 0.6127 0.6853 0.8278
No log 9.5417 458 0.7182 0.5555 0.7182 0.8475
No log 9.5833 460 0.6705 0.6377 0.6705 0.8188
No log 9.625 462 0.6488 0.5529 0.6488 0.8055
No log 9.6667 464 0.6706 0.5704 0.6706 0.8189
No log 9.7083 466 0.6649 0.5704 0.6649 0.8154
No log 9.75 468 0.6533 0.5713 0.6533 0.8083
No log 9.7917 470 0.6769 0.6429 0.6769 0.8227
No log 9.8333 472 0.6660 0.6677 0.6660 0.8161
No log 9.875 474 0.6631 0.6314 0.6631 0.8143
No log 9.9167 476 0.6789 0.6282 0.6789 0.8240
No log 9.9583 478 0.6889 0.4981 0.6889 0.8300
No log 10.0 480 0.7168 0.4839 0.7168 0.8466
No log 10.0417 482 0.7317 0.4843 0.7317 0.8554
No log 10.0833 484 0.7031 0.5495 0.7031 0.8385
No log 10.125 486 0.6699 0.4659 0.6699 0.8185
No log 10.1667 488 0.6566 0.5174 0.6566 0.8103
No log 10.2083 490 0.6678 0.5467 0.6678 0.8172
No log 10.25 492 0.7801 0.5526 0.7801 0.8832
No log 10.2917 494 0.8427 0.5526 0.8427 0.9180
No log 10.3333 496 0.7703 0.5526 0.7703 0.8777
No log 10.375 498 0.6723 0.4898 0.6723 0.8199
0.318 10.4167 500 0.6825 0.5568 0.6825 0.8261
0.318 10.4583 502 0.6821 0.5468 0.6821 0.8259
0.318 10.5 504 0.6591 0.5183 0.6591 0.8118
0.318 10.5417 506 0.6899 0.4610 0.6899 0.8306
0.318 10.5833 508 0.7232 0.4958 0.7232 0.8504
0.318 10.625 510 0.7086 0.4473 0.7086 0.8418

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k15_task5_organization

Finetuned
(4019)
this model