ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k18_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7797
  • Qwk: 0.6475
  • Mse: 0.7797
  • Rmse: 0.8830

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0235 2 7.1513 0.0057 7.1513 2.6742
No log 0.0471 4 5.3367 0.0361 5.3367 2.3101
No log 0.0706 6 4.5984 -0.0686 4.5984 2.1444
No log 0.0941 8 3.8128 -0.0652 3.8128 1.9526
No log 0.1176 10 2.5825 0.0149 2.5825 1.6070
No log 0.1412 12 1.9712 0.1217 1.9712 1.4040
No log 0.1647 14 2.2060 0.1260 2.2060 1.4852
No log 0.1882 16 3.3240 0.0223 3.3240 1.8232
No log 0.2118 18 3.2154 0.0331 3.2154 1.7932
No log 0.2353 20 2.1453 0.1679 2.1453 1.4647
No log 0.2588 22 1.7201 0.1835 1.7201 1.3115
No log 0.2824 24 2.2789 -0.0351 2.2789 1.5096
No log 0.3059 26 2.4309 -0.1197 2.4309 1.5591
No log 0.3294 28 2.2322 -0.0348 2.2322 1.4941
No log 0.3529 30 1.8203 0.1667 1.8203 1.3492
No log 0.3765 32 1.8644 0.2586 1.8644 1.3654
No log 0.4 34 2.1622 0.1931 2.1622 1.4704
No log 0.4235 36 2.3892 0.1046 2.3892 1.5457
No log 0.4471 38 2.2372 0.1351 2.2372 1.4957
No log 0.4706 40 1.9280 0.3200 1.9280 1.3885
No log 0.4941 42 1.8649 0.3193 1.8649 1.3656
No log 0.5176 44 1.8039 0.3220 1.8039 1.3431
No log 0.5412 46 1.7254 0.2783 1.7254 1.3135
No log 0.5647 48 1.6972 0.384 1.6972 1.3027
No log 0.5882 50 1.7927 0.3913 1.7927 1.3389
No log 0.6118 52 2.0378 0.3444 2.0378 1.4275
No log 0.6353 54 2.4412 0.3243 2.4412 1.5624
No log 0.6588 56 2.2854 0.3444 2.2854 1.5118
No log 0.6824 58 1.5835 0.5101 1.5835 1.2584
No log 0.7059 60 1.3339 0.4531 1.3339 1.1549
No log 0.7294 62 1.4547 0.4409 1.4547 1.2061
No log 0.7529 64 1.5052 0.4480 1.5052 1.2269
No log 0.7765 66 1.6058 0.4409 1.6058 1.2672
No log 0.8 68 1.7134 0.4 1.7134 1.3090
No log 0.8235 70 1.7073 0.3780 1.7073 1.3066
No log 0.8471 72 1.6036 0.4094 1.6036 1.2663
No log 0.8706 74 1.4113 0.4160 1.4113 1.1880
No log 0.8941 76 1.3610 0.4296 1.3610 1.1666
No log 0.9176 78 1.8110 0.3467 1.8110 1.3457
No log 0.9412 80 2.3152 0.3721 2.3152 1.5216
No log 0.9647 82 2.1450 0.3721 2.1450 1.4646
No log 0.9882 84 2.1453 0.4181 2.1453 1.4647
No log 1.0118 86 1.9027 0.4678 1.9027 1.3794
No log 1.0353 88 1.4828 0.4932 1.4828 1.2177
No log 1.0588 90 1.2607 0.3968 1.2607 1.1228
No log 1.0824 92 1.2061 0.3559 1.2061 1.0982
No log 1.1059 94 1.2265 0.3697 1.2265 1.1075
No log 1.1294 96 1.3634 0.4480 1.3634 1.1676
No log 1.1529 98 1.7185 0.4054 1.7185 1.3109
No log 1.1765 100 2.0706 0.3721 2.0706 1.4390
No log 1.2 102 2.0089 0.3955 2.0089 1.4174
No log 1.2235 104 1.5145 0.5235 1.5145 1.2307
No log 1.2471 106 1.2294 0.5676 1.2294 1.1088
No log 1.2706 108 1.0633 0.5426 1.0633 1.0311
No log 1.2941 110 0.9779 0.5116 0.9779 0.9889
No log 1.3176 112 0.9831 0.512 0.9831 0.9915
No log 1.3412 114 1.0765 0.5156 1.0765 1.0376
No log 1.3647 116 1.2660 0.4697 1.2660 1.1251
No log 1.3882 118 1.2835 0.4697 1.2835 1.1329
No log 1.4118 120 1.1538 0.4677 1.1538 1.0742
No log 1.4353 122 1.0555 0.4667 1.0555 1.0274
No log 1.4588 124 1.0310 0.4793 1.0310 1.0154
No log 1.4824 126 1.0326 0.4878 1.0326 1.0162
No log 1.5059 128 1.0176 0.5041 1.0176 1.0088
No log 1.5294 130 1.0302 0.5496 1.0302 1.0150
No log 1.5529 132 1.0661 0.6043 1.0661 1.0325
No log 1.5765 134 1.0133 0.6014 1.0133 1.0066
No log 1.6 136 0.9556 0.6294 0.9556 0.9775
No log 1.6235 138 0.8757 0.6712 0.8757 0.9358
No log 1.6471 140 0.9208 0.6438 0.9208 0.9596
No log 1.6706 142 1.0086 0.5960 1.0086 1.0043
No log 1.6941 144 1.1046 0.5638 1.1046 1.0510
No log 1.7176 146 1.1986 0.5379 1.1986 1.0948
No log 1.7412 148 1.1357 0.5547 1.1357 1.0657
No log 1.7647 150 0.9825 0.5899 0.9825 0.9912
No log 1.7882 152 0.9008 0.6111 0.9008 0.9491
No log 1.8118 154 0.8493 0.6901 0.8493 0.9216
No log 1.8353 156 0.9154 0.7133 0.9154 0.9567
No log 1.8588 158 0.8761 0.7347 0.8761 0.9360
No log 1.8824 160 0.8744 0.7347 0.8744 0.9351
No log 1.9059 162 0.8585 0.7347 0.8585 0.9265
No log 1.9294 164 0.9820 0.5455 0.9820 0.9910
No log 1.9529 166 0.9829 0.5563 0.9829 0.9914
No log 1.9765 168 0.7933 0.7564 0.7933 0.8907
No log 2.0 170 0.7615 0.76 0.7615 0.8726
No log 2.0235 172 0.7041 0.7662 0.7041 0.8391
No log 2.0471 174 0.7323 0.7468 0.7323 0.8558
No log 2.0706 176 0.8193 0.7453 0.8193 0.9051
No log 2.0941 178 0.8387 0.6800 0.8387 0.9158
No log 2.1176 180 0.8815 0.6573 0.8815 0.9389
No log 2.1412 182 0.9318 0.6759 0.9318 0.9653
No log 2.1647 184 0.9327 0.6939 0.9327 0.9658
No log 2.1882 186 0.9476 0.6621 0.9476 0.9735
No log 2.2118 188 0.9416 0.6621 0.9416 0.9704
No log 2.2353 190 0.8936 0.6846 0.8936 0.9453
No log 2.2588 192 0.8687 0.6846 0.8687 0.9320
No log 2.2824 194 0.8657 0.6667 0.8657 0.9304
No log 2.3059 196 0.8989 0.6483 0.8989 0.9481
No log 2.3294 198 0.9758 0.5481 0.9758 0.9878
No log 2.3529 200 1.0116 0.5649 1.0116 1.0058
No log 2.3765 202 1.0365 0.5271 1.0365 1.0181
No log 2.4 204 0.9755 0.6222 0.9755 0.9877
No log 2.4235 206 0.8460 0.6571 0.8460 0.9198
No log 2.4471 208 0.7986 0.6950 0.7986 0.8937
No log 2.4706 210 0.7332 0.7347 0.7332 0.8563
No log 2.4941 212 0.7424 0.7237 0.7424 0.8616
No log 2.5176 214 0.7494 0.7320 0.7494 0.8657
No log 2.5412 216 0.7633 0.7162 0.7633 0.8737
No log 2.5647 218 0.7755 0.6897 0.7755 0.8807
No log 2.5882 220 0.8203 0.5797 0.8203 0.9057
No log 2.6118 222 0.9501 0.6187 0.9501 0.9747
No log 2.6353 224 0.9565 0.6338 0.9565 0.9780
No log 2.6588 226 0.9130 0.6294 0.9130 0.9555
No log 2.6824 228 1.0359 0.6483 1.0359 1.0178
No log 2.7059 230 1.0342 0.6575 1.0342 1.0169
No log 2.7294 232 0.8806 0.6483 0.8806 0.9384
No log 2.7529 234 0.7817 0.6187 0.7817 0.8841
No log 2.7765 236 0.8136 0.5714 0.8136 0.9020
No log 2.8 238 0.8229 0.6471 0.8229 0.9072
No log 2.8235 240 0.8488 0.5970 0.8488 0.9213
No log 2.8471 242 0.8505 0.6389 0.8505 0.9222
No log 2.8706 244 0.8268 0.6757 0.8268 0.9093
No log 2.8941 246 0.7546 0.7632 0.7546 0.8687
No log 2.9176 248 0.7712 0.7172 0.7712 0.8782
No log 2.9412 250 0.8888 0.6861 0.8888 0.9428
No log 2.9647 252 0.9721 0.6475 0.9721 0.9859
No log 2.9882 254 0.9414 0.6861 0.9414 0.9703
No log 3.0118 256 0.8557 0.7206 0.8557 0.9250
No log 3.0353 258 0.8092 0.5926 0.8092 0.8995
No log 3.0588 260 0.8051 0.6434 0.8051 0.8973
No log 3.0824 262 0.7341 0.6939 0.7341 0.8568
No log 3.1059 264 0.7087 0.7297 0.7087 0.8418
No log 3.1294 266 0.7599 0.6912 0.7599 0.8717
No log 3.1529 268 0.7643 0.6812 0.7643 0.8742
No log 3.1765 270 0.7664 0.6212 0.7664 0.8755
No log 3.2 272 0.8805 0.5797 0.8805 0.9384
No log 3.2235 274 0.9298 0.6029 0.9298 0.9642
No log 3.2471 276 0.8724 0.5469 0.8724 0.9340
No log 3.2706 278 0.8347 0.6466 0.8347 0.9136
No log 3.2941 280 0.8774 0.6364 0.8774 0.9367
No log 3.3176 282 0.8663 0.6716 0.8663 0.9308
No log 3.3412 284 0.8306 0.6212 0.8306 0.9114
No log 3.3647 286 0.8276 0.5802 0.8276 0.9097
No log 3.3882 288 0.8473 0.6479 0.8473 0.9205
No log 3.4118 290 0.7691 0.7297 0.7691 0.8770
No log 3.4353 292 0.7098 0.6901 0.7098 0.8425
No log 3.4588 294 0.8517 0.6418 0.8517 0.9229
No log 3.4824 296 1.0941 0.5077 1.0941 1.0460
No log 3.5059 298 1.5520 0.4203 1.5520 1.2458
No log 3.5294 300 1.6862 0.4490 1.6862 1.2985
No log 3.5529 302 1.5918 0.4932 1.5918 1.2617
No log 3.5765 304 1.1378 0.4769 1.1378 1.0667
No log 3.6 306 0.8575 0.5954 0.8575 0.9260
No log 3.6235 308 0.8130 0.6963 0.8130 0.9017
No log 3.6471 310 0.9621 0.6154 0.9621 0.9808
No log 3.6706 312 1.1487 0.6241 1.1487 1.0718
No log 3.6941 314 1.1207 0.5909 1.1207 1.0586
No log 3.7176 316 1.0035 0.6212 1.0035 1.0017
No log 3.7412 318 1.0275 0.5736 1.0275 1.0136
No log 3.7647 320 0.9980 0.5736 0.9980 0.9990
No log 3.7882 322 0.9456 0.5954 0.9456 0.9724
No log 3.8118 324 0.8987 0.6412 0.8987 0.9480
No log 3.8353 326 0.8251 0.6812 0.8251 0.9083
No log 3.8588 328 0.7546 0.7286 0.7546 0.8687
No log 3.8824 330 0.7546 0.6950 0.7546 0.8687
No log 3.9059 332 0.7405 0.6857 0.7405 0.8605
No log 3.9294 334 0.7417 0.6715 0.7417 0.8612
No log 3.9529 336 0.7859 0.6715 0.7859 0.8865
No log 3.9765 338 0.7744 0.7034 0.7744 0.8800
No log 4.0 340 0.7701 0.7368 0.7701 0.8775
No log 4.0235 342 0.6813 0.7712 0.6813 0.8254
No log 4.0471 344 0.6354 0.8354 0.6354 0.7971
No log 4.0706 346 0.6388 0.7929 0.6388 0.7993
No log 4.0941 348 0.6374 0.8049 0.6374 0.7984
No log 4.1176 350 0.6374 0.7733 0.6374 0.7984
No log 4.1412 352 0.6911 0.6763 0.6911 0.8313
No log 4.1647 354 0.7357 0.6763 0.7357 0.8577
No log 4.1882 356 0.7662 0.7246 0.7662 0.8753
No log 4.2118 358 0.8056 0.6119 0.8056 0.8975
No log 4.2353 360 0.8207 0.6165 0.8207 0.9059
No log 4.2588 362 0.8123 0.6061 0.8123 0.9013
No log 4.2824 364 0.7934 0.6569 0.7934 0.8907
No log 4.3059 366 0.7780 0.6569 0.7780 0.8821
No log 4.3294 368 0.7865 0.6569 0.7865 0.8869
No log 4.3529 370 0.7835 0.6571 0.7835 0.8852
No log 4.3765 372 0.7384 0.7034 0.7384 0.8593
No log 4.4 374 0.6885 0.7413 0.6885 0.8297
No log 4.4235 376 0.7346 0.7143 0.7346 0.8571
No log 4.4471 378 0.7526 0.7143 0.7526 0.8675
No log 4.4706 380 0.6985 0.6993 0.6985 0.8358
No log 4.4941 382 0.6957 0.6950 0.6957 0.8341
No log 4.5176 384 0.7786 0.7237 0.7786 0.8824
No log 4.5412 386 0.7896 0.7114 0.7896 0.8886
No log 4.5647 388 0.7601 0.6812 0.7601 0.8718
No log 4.5882 390 0.7714 0.6667 0.7714 0.8783
No log 4.6118 392 0.7650 0.6565 0.7650 0.8747
No log 4.6353 394 0.7773 0.6370 0.7773 0.8817
No log 4.6588 396 0.8296 0.6324 0.8296 0.9108
No log 4.6824 398 0.8847 0.6074 0.8847 0.9406
No log 4.7059 400 0.8587 0.5970 0.8587 0.9267
No log 4.7294 402 0.8329 0.6260 0.8329 0.9126
No log 4.7529 404 0.8014 0.6212 0.8014 0.8952
No log 4.7765 406 0.7857 0.6107 0.7857 0.8864
No log 4.8 408 0.7935 0.5556 0.7935 0.8908
No log 4.8235 410 0.7997 0.576 0.7997 0.8943
No log 4.8471 412 0.7840 0.625 0.7840 0.8854
No log 4.8706 414 0.7841 0.6032 0.7841 0.8855
No log 4.8941 416 0.7578 0.6718 0.7578 0.8705
No log 4.9176 418 0.7579 0.6767 0.7579 0.8706
No log 4.9412 420 0.8337 0.6617 0.8337 0.9131
No log 4.9647 422 0.8584 0.6617 0.8584 0.9265
No log 4.9882 424 0.7862 0.6963 0.7862 0.8867
No log 5.0118 426 0.7435 0.6515 0.7435 0.8622
No log 5.0353 428 0.7718 0.6462 0.7718 0.8785
No log 5.0588 430 0.8246 0.5846 0.8246 0.9081
No log 5.0824 432 0.8891 0.6 0.8891 0.9429
No log 5.1059 434 0.8991 0.6 0.8991 0.9482
No log 5.1294 436 0.8759 0.6047 0.8759 0.9359
No log 5.1529 438 0.8288 0.6154 0.8288 0.9104
No log 5.1765 440 0.8130 0.6713 0.8130 0.9017
No log 5.2 442 0.7814 0.7105 0.7814 0.8839
No log 5.2235 444 0.7240 0.7226 0.7240 0.8509
No log 5.2471 446 0.6650 0.7417 0.6650 0.8155
No log 5.2706 448 0.6624 0.7050 0.6624 0.8139
No log 5.2941 450 0.6779 0.6957 0.6779 0.8234
No log 5.3176 452 0.7125 0.7059 0.7125 0.8441
No log 5.3412 454 0.7424 0.7059 0.7424 0.8617
No log 5.3647 456 0.7862 0.6107 0.7862 0.8867
No log 5.3882 458 0.8102 0.6 0.8102 0.9001
No log 5.4118 460 0.7951 0.6364 0.7951 0.8917
No log 5.4353 462 0.7672 0.6815 0.7672 0.8759
No log 5.4588 464 0.7633 0.7299 0.7633 0.8737
No log 5.4824 466 0.7415 0.7153 0.7415 0.8611
No log 5.5059 468 0.7357 0.6316 0.7357 0.8577
No log 5.5294 470 0.7648 0.6812 0.7648 0.8745
No log 5.5529 472 0.7494 0.6812 0.7494 0.8657
No log 5.5765 474 0.7006 0.6617 0.7006 0.8370
No log 5.6 476 0.7104 0.7059 0.7104 0.8428
No log 5.6235 478 0.7093 0.7153 0.7093 0.8422
No log 5.6471 480 0.7061 0.7445 0.7061 0.8403
No log 5.6706 482 0.7499 0.7007 0.7499 0.8660
No log 5.6941 484 0.7725 0.7101 0.7725 0.8789
No log 5.7176 486 0.8001 0.6912 0.8001 0.8945
No log 5.7412 488 0.8150 0.6165 0.8150 0.9028
No log 5.7647 490 0.7796 0.6316 0.7796 0.8829
No log 5.7882 492 0.7847 0.6667 0.7847 0.8858
No log 5.8118 494 0.7907 0.6716 0.7907 0.8892
No log 5.8353 496 0.8036 0.6515 0.8036 0.8965
No log 5.8588 498 0.8203 0.6515 0.8203 0.9057
0.4503 5.8824 500 0.8210 0.6767 0.8210 0.9061
0.4503 5.9059 502 0.8295 0.5954 0.8295 0.9108
0.4503 5.9294 504 0.8986 0.6119 0.8986 0.9480
0.4503 5.9529 506 0.9407 0.6286 0.9407 0.9699
0.4503 5.9765 508 0.8855 0.6620 0.8855 0.9410
0.4503 6.0 510 0.7797 0.6475 0.7797 0.8830

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k18_task1_organization

Finetuned
(4019)
this model