ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k17_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9987
  • Qwk: 0.5985
  • Mse: 0.9987
  • Rmse: 0.9994

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.025 2 6.9425 0.0116 6.9425 2.6349
No log 0.05 4 4.6678 0.0791 4.6678 2.1605
No log 0.075 6 3.7367 -0.0410 3.7367 1.9331
No log 0.1 8 2.5959 0.1667 2.5959 1.6112
No log 0.125 10 2.0113 0.1311 2.0113 1.4182
No log 0.15 12 1.7662 0.2735 1.7662 1.3290
No log 0.175 14 2.4266 0.1176 2.4266 1.5578
No log 0.2 16 2.4705 0.1169 2.4705 1.5718
No log 0.225 18 1.9987 0.1417 1.9987 1.4137
No log 0.25 20 1.7881 0.1982 1.7881 1.3372
No log 0.275 22 1.6996 0.1698 1.6996 1.3037
No log 0.3 24 1.6731 0.1524 1.6731 1.2935
No log 0.325 26 1.7013 0.1682 1.7013 1.3043
No log 0.35 28 1.7568 0.2810 1.7568 1.3255
No log 0.375 30 2.4170 0.0552 2.4170 1.5547
No log 0.4 32 2.8322 0.0135 2.8322 1.6829
No log 0.425 34 2.6997 0.0408 2.6997 1.6431
No log 0.45 36 2.2132 0.2014 2.2132 1.4877
No log 0.475 38 1.5860 0.3802 1.5860 1.2594
No log 0.5 40 1.4381 0.2832 1.4381 1.1992
No log 0.525 42 1.5351 0.2727 1.5351 1.2390
No log 0.55 44 1.5370 0.2832 1.5370 1.2398
No log 0.575 46 1.6090 0.2810 1.6090 1.2685
No log 0.6 48 2.0602 0.2137 2.0602 1.4353
No log 0.625 50 2.3454 0.1241 2.3454 1.5315
No log 0.65 52 2.2931 0.1325 2.2931 1.5143
No log 0.675 54 2.1929 0.1806 2.1929 1.4809
No log 0.7 56 2.2861 0.1935 2.2861 1.5120
No log 0.725 58 2.4447 0.2105 2.4447 1.5636
No log 0.75 60 1.9911 0.3418 1.9911 1.4110
No log 0.775 62 1.4111 0.4861 1.4111 1.1879
No log 0.8 64 1.2572 0.5101 1.2572 1.1212
No log 0.825 66 1.2415 0.5548 1.2415 1.1142
No log 0.85 68 1.0526 0.6164 1.0526 1.0260
No log 0.875 70 0.9509 0.6710 0.9509 0.9751
No log 0.9 72 0.9725 0.6625 0.9725 0.9862
No log 0.925 74 1.0715 0.6587 1.0715 1.0351
No log 0.95 76 1.0800 0.6708 1.0800 1.0392
No log 0.975 78 1.2099 0.6265 1.2099 1.0999
No log 1.0 80 1.4465 0.5244 1.4465 1.2027
No log 1.025 82 2.0395 0.3913 2.0395 1.4281
No log 1.05 84 2.2266 0.3608 2.2266 1.4922
No log 1.075 86 1.3922 0.5989 1.3922 1.1799
No log 1.1 88 0.9374 0.6420 0.9374 0.9682
No log 1.125 90 0.9331 0.6538 0.9331 0.9660
No log 1.15 92 1.0434 0.5912 1.0434 1.0215
No log 1.175 94 1.1416 0.6433 1.1416 1.0685
No log 1.2 96 1.2508 0.6437 1.2508 1.1184
No log 1.225 98 1.0976 0.6386 1.0976 1.0477
No log 1.25 100 0.8989 0.6447 0.8989 0.9481
No log 1.275 102 0.8631 0.6528 0.8631 0.9290
No log 1.3 104 0.9035 0.6197 0.9035 0.9505
No log 1.325 106 0.9881 0.6479 0.9881 0.9940
No log 1.35 108 1.2131 0.6323 1.2131 1.1014
No log 1.375 110 1.1882 0.6329 1.1882 1.0900
No log 1.4 112 1.2679 0.5963 1.2679 1.1260
No log 1.425 114 0.8841 0.7237 0.8841 0.9403
No log 1.45 116 0.7040 0.7027 0.7040 0.8391
No log 1.475 118 0.7067 0.7027 0.7067 0.8406
No log 1.5 120 0.9209 0.7089 0.9209 0.9596
No log 1.525 122 1.0777 0.6497 1.0777 1.0381
No log 1.55 124 0.9210 0.6803 0.9210 0.9597
No log 1.575 126 0.8394 0.6197 0.8394 0.9162
No log 1.6 128 0.8693 0.6197 0.8693 0.9324
No log 1.625 130 1.0485 0.6503 1.0485 1.0240
No log 1.65 132 1.3354 0.5989 1.3354 1.1556
No log 1.675 134 1.2366 0.6215 1.2366 1.1120
No log 1.7 136 0.9934 0.6296 0.9934 0.9967
No log 1.725 138 0.8984 0.6846 0.8984 0.9478
No log 1.75 140 0.9054 0.6622 0.9054 0.9515
No log 1.775 142 0.8642 0.6622 0.8642 0.9296
No log 1.8 144 0.8893 0.6667 0.8893 0.9430
No log 1.825 146 0.9969 0.6434 0.9969 0.9985
No log 1.85 148 1.1171 0.6197 1.1171 1.0569
No log 1.875 150 1.1893 0.6164 1.1893 1.0905
No log 1.9 152 1.0438 0.6154 1.0438 1.0216
No log 1.925 154 0.9360 0.6165 0.9360 0.9675
No log 1.95 156 0.8769 0.6277 0.8769 0.9364
No log 1.975 158 0.8469 0.6475 0.8469 0.9203
No log 2.0 160 0.8444 0.6619 0.8444 0.9189
No log 2.025 162 0.9125 0.6619 0.9125 0.9552
No log 2.05 164 0.9600 0.6316 0.9600 0.9798
No log 2.075 166 0.8937 0.6519 0.8937 0.9453
No log 2.1 168 0.9285 0.6301 0.9285 0.9636
No log 2.125 170 1.1405 0.6447 1.1405 1.0679
No log 2.15 172 1.6078 0.5087 1.6078 1.2680
No log 2.175 174 1.7665 0.4659 1.7665 1.3291
No log 2.2 176 1.5158 0.5125 1.5158 1.2312
No log 2.225 178 1.1418 0.6164 1.1418 1.0685
No log 2.25 180 0.8544 0.6429 0.8544 0.9244
No log 2.275 182 0.8031 0.6765 0.8031 0.8962
No log 2.3 184 0.7816 0.7222 0.7816 0.8841
No log 2.325 186 0.9507 0.6531 0.9507 0.9750
No log 2.35 188 1.1301 0.5769 1.1301 1.0631
No log 2.375 190 1.2377 0.6127 1.2377 1.1125
No log 2.4 192 0.9863 0.6667 0.9863 0.9931
No log 2.425 194 0.7540 0.7432 0.7540 0.8683
No log 2.45 196 0.8149 0.7101 0.8149 0.9027
No log 2.475 198 0.8841 0.6569 0.8841 0.9403
No log 2.5 200 0.9671 0.5312 0.9671 0.9834
No log 2.525 202 0.9867 0.5606 0.9867 0.9933
No log 2.55 204 1.0722 0.5882 1.0722 1.0355
No log 2.575 206 1.1045 0.5362 1.1045 1.0510
No log 2.6 208 1.0224 0.5588 1.0224 1.0112
No log 2.625 210 1.0153 0.5496 1.0153 1.0076
No log 2.65 212 0.9831 0.5564 0.9831 0.9915
No log 2.675 214 0.8337 0.6667 0.8337 0.9131
No log 2.7 216 0.8534 0.6000 0.8534 0.9238
No log 2.725 218 0.9207 0.6099 0.9207 0.9595
No log 2.75 220 0.9914 0.6538 0.9914 0.9957
No log 2.775 222 1.0691 0.6303 1.0691 1.0340
No log 2.8 224 0.9631 0.6545 0.9631 0.9814
No log 2.825 226 0.9389 0.6705 0.9389 0.9690
No log 2.85 228 0.8474 0.7024 0.8474 0.9205
No log 2.875 230 0.7313 0.7179 0.7313 0.8552
No log 2.9 232 0.7862 0.6974 0.7862 0.8867
No log 2.925 234 0.9858 0.6918 0.9858 0.9929
No log 2.95 236 1.4576 0.5889 1.4576 1.2073
No log 2.975 238 1.8843 0.4592 1.8843 1.3727
No log 3.0 240 1.9333 0.4523 1.9333 1.3904
No log 3.025 242 1.6463 0.5185 1.6463 1.2831
No log 3.05 244 1.1521 0.6335 1.1521 1.0734
No log 3.075 246 0.8810 0.6575 0.8810 0.9386
No log 3.1 248 0.8471 0.6759 0.8471 0.9204
No log 3.125 250 0.8482 0.6892 0.8482 0.9210
No log 3.15 252 0.8537 0.6577 0.8537 0.9240
No log 3.175 254 0.8320 0.6667 0.8320 0.9121
No log 3.2 256 0.8531 0.6345 0.8531 0.9236
No log 3.225 258 0.8499 0.6351 0.8499 0.9219
No log 3.25 260 0.9221 0.6538 0.9221 0.9602
No log 3.275 262 0.9633 0.6497 0.9633 0.9815
No log 3.3 264 0.8518 0.6301 0.8518 0.9229
No log 3.325 266 0.8176 0.6099 0.8176 0.9042
No log 3.35 268 0.8321 0.6 0.8321 0.9122
No log 3.375 270 0.9192 0.6752 0.9192 0.9587
No log 3.4 272 0.8709 0.6752 0.8709 0.9332
No log 3.425 274 0.7867 0.6918 0.7867 0.8870
No log 3.45 276 0.7338 0.7375 0.7338 0.8566
No log 3.475 278 0.8213 0.7 0.8213 0.9063
No log 3.5 280 0.7985 0.6667 0.7985 0.8936
No log 3.525 282 0.8035 0.6711 0.8035 0.8964
No log 3.55 284 0.8776 0.6338 0.8776 0.9368
No log 3.575 286 0.9681 0.6197 0.9681 0.9839
No log 3.6 288 0.9940 0.6099 0.9940 0.9970
No log 3.625 290 1.0078 0.6087 1.0078 1.0039
No log 3.65 292 1.0138 0.6286 1.0138 1.0069
No log 3.675 294 0.9808 0.5839 0.9808 0.9904
No log 3.7 296 0.8919 0.5926 0.8919 0.9444
No log 3.725 298 0.8304 0.5926 0.8304 0.9113
No log 3.75 300 0.7866 0.6131 0.7866 0.8869
No log 3.775 302 0.7936 0.6029 0.7936 0.8909
No log 3.8 304 0.8157 0.5926 0.8157 0.9032
No log 3.825 306 0.8493 0.6522 0.8493 0.9216
No log 3.85 308 0.8233 0.7 0.8233 0.9074
No log 3.875 310 0.7657 0.7 0.7657 0.8751
No log 3.9 312 0.8481 0.6486 0.8481 0.9209
No log 3.925 314 0.9011 0.6316 0.9011 0.9493
No log 3.95 316 0.8544 0.6056 0.8544 0.9243
No log 3.975 318 0.9198 0.6056 0.9198 0.9591
No log 4.0 320 1.1300 0.5960 1.1300 1.0630
No log 4.025 322 1.2591 0.6335 1.2591 1.1221
No log 4.05 324 1.2526 0.6071 1.2526 1.1192
No log 4.075 326 0.9381 0.6410 0.9381 0.9686
No log 4.1 328 0.6628 0.7517 0.6628 0.8141
No log 4.125 330 0.6455 0.7660 0.6455 0.8034
No log 4.15 332 0.6457 0.7778 0.6457 0.8036
No log 4.175 334 0.6776 0.7632 0.6776 0.8232
No log 4.2 336 0.7910 0.6839 0.7910 0.8894
No log 4.225 338 0.9016 0.6316 0.9016 0.9495
No log 4.25 340 0.8816 0.6301 0.8816 0.9389
No log 4.275 342 0.8483 0.6099 0.8483 0.9210
No log 4.3 344 0.8297 0.6713 0.8297 0.9109
No log 4.325 346 0.7817 0.7034 0.7817 0.8841
No log 4.35 348 0.7551 0.7297 0.7551 0.8690
No log 4.375 350 0.7752 0.6928 0.7752 0.8804
No log 4.4 352 0.7691 0.6711 0.7691 0.8770
No log 4.425 354 0.7098 0.7673 0.7098 0.8425
No log 4.45 356 0.6837 0.7682 0.6837 0.8269
No log 4.475 358 0.6989 0.7733 0.6989 0.8360
No log 4.5 360 0.7849 0.6667 0.7849 0.8859
No log 4.525 362 0.8610 0.6301 0.8610 0.9279
No log 4.55 364 0.8408 0.6043 0.8408 0.9170
No log 4.575 366 0.8340 0.6286 0.8340 0.9132
No log 4.6 368 0.8531 0.6143 0.8531 0.9236
No log 4.625 370 0.8540 0.6043 0.8540 0.9241
No log 4.65 372 0.8784 0.6043 0.8784 0.9372
No log 4.675 374 0.9352 0.6207 0.9352 0.9670
No log 4.7 376 0.8653 0.6294 0.8653 0.9302
No log 4.725 378 0.7789 0.6761 0.7789 0.8826
No log 4.75 380 0.7512 0.7465 0.7512 0.8667
No log 4.775 382 0.7818 0.7092 0.7818 0.8842
No log 4.8 384 0.8804 0.5758 0.8804 0.9383
No log 4.825 386 1.0213 0.5693 1.0213 1.0106
No log 4.85 388 1.1198 0.5755 1.1198 1.0582
No log 4.875 390 1.0920 0.6014 1.0920 1.0450
No log 4.9 392 0.9692 0.6164 0.9692 0.9845
No log 4.925 394 0.7519 0.7368 0.7519 0.8671
No log 4.95 396 0.6612 0.7417 0.6612 0.8132
No log 4.975 398 0.6384 0.7974 0.6384 0.7990
No log 5.0 400 0.6597 0.7333 0.6597 0.8122
No log 5.025 402 0.6749 0.7417 0.6749 0.8215
No log 5.05 404 0.7219 0.7133 0.7219 0.8497
No log 5.075 406 0.8011 0.6667 0.8011 0.8951
No log 5.1 408 0.8680 0.6331 0.8680 0.9316
No log 5.125 410 0.8584 0.6232 0.8584 0.9265
No log 5.15 412 0.7960 0.6619 0.7960 0.8922
No log 5.175 414 0.7888 0.6619 0.7888 0.8882
No log 5.2 416 0.8307 0.6087 0.8307 0.9115
No log 5.225 418 0.8231 0.6029 0.8231 0.9073
No log 5.25 420 0.8150 0.6857 0.8150 0.9028
No log 5.275 422 0.8499 0.6269 0.8499 0.9219
No log 5.3 424 0.8638 0.5606 0.8638 0.9294
No log 5.325 426 0.9075 0.5778 0.9075 0.9526
No log 5.35 428 1.0215 0.6351 1.0215 1.0107
No log 5.375 430 0.9716 0.6490 0.9716 0.9857
No log 5.4 432 0.8759 0.6301 0.8759 0.9359
No log 5.425 434 0.8203 0.6056 0.8203 0.9057
No log 5.45 436 0.7731 0.6712 0.7731 0.8793
No log 5.475 438 0.8191 0.5957 0.8191 0.9051
No log 5.5 440 0.9872 0.6490 0.9872 0.9936
No log 5.525 442 1.0645 0.6490 1.0645 1.0317
No log 5.55 444 0.9711 0.5985 0.9711 0.9854
No log 5.575 446 0.8548 0.5426 0.8548 0.9246
No log 5.6 448 0.8236 0.6519 0.8236 0.9075
No log 5.625 450 0.8239 0.6618 0.8239 0.9077
No log 5.65 452 0.8340 0.6618 0.8340 0.9132
No log 5.675 454 0.7985 0.6857 0.7985 0.8936
No log 5.7 456 0.7770 0.7237 0.7770 0.8815
No log 5.725 458 0.7540 0.6944 0.7540 0.8683
No log 5.75 460 0.7782 0.6519 0.7782 0.8821
No log 5.775 462 0.8431 0.6212 0.8431 0.9182
No log 5.8 464 0.9894 0.5714 0.9894 0.9947
No log 5.825 466 1.1158 0.5926 1.1158 1.0563
No log 5.85 468 1.1052 0.5926 1.1052 1.0513
No log 5.875 470 0.9579 0.5821 0.9579 0.9787
No log 5.9 472 0.8371 0.6047 0.8371 0.9149
No log 5.925 474 0.8352 0.6047 0.8352 0.9139
No log 5.95 476 0.8350 0.6418 0.8350 0.9138
No log 5.975 478 0.8388 0.6316 0.8388 0.9158
No log 6.0 480 0.8951 0.5714 0.8951 0.9461
No log 6.025 482 0.9150 0.5839 0.9150 0.9566
No log 6.05 484 0.8264 0.6370 0.8264 0.9091
No log 6.075 486 0.8384 0.6119 0.8384 0.9156
No log 6.1 488 0.9377 0.6 0.9377 0.9684
No log 6.125 490 1.0566 0.6335 1.0566 1.0279
No log 6.15 492 1.0038 0.6341 1.0038 1.0019
No log 6.175 494 1.0560 0.6506 1.0560 1.0276
No log 6.2 496 0.9637 0.6747 0.9637 0.9817
No log 6.225 498 0.7791 0.6803 0.7791 0.8827
0.4372 6.25 500 0.7123 0.7234 0.7123 0.8440
0.4372 6.275 502 0.7285 0.7246 0.7285 0.8535
0.4372 6.3 504 0.7426 0.6861 0.7426 0.8617
0.4372 6.325 506 0.8015 0.6423 0.8015 0.8953
0.4372 6.35 508 0.9673 0.6241 0.9673 0.9835
0.4372 6.375 510 1.0598 0.6 1.0598 1.0294
0.4372 6.4 512 0.9987 0.5985 0.9987 0.9994

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k17_task1_organization

Finetuned
(4019)
this model