ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k10_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7442
  • Qwk: 0.7259
  • Mse: 0.7442
  • Rmse: 0.8627

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0417 2 6.8112 0.0179 6.8112 2.6098
No log 0.0833 4 4.3106 0.0750 4.3106 2.0762
No log 0.125 6 2.9553 0.0366 2.9553 1.7191
No log 0.1667 8 2.3593 0.0735 2.3593 1.5360
No log 0.2083 10 2.0630 0.1148 2.0630 1.4363
No log 0.25 12 1.9345 0.1356 1.9345 1.3909
No log 0.2917 14 1.8193 0.1053 1.8193 1.3488
No log 0.3333 16 1.9656 0.2308 1.9656 1.4020
No log 0.375 18 3.2403 0.1250 3.2402 1.8001
No log 0.4167 20 5.4345 0.0139 5.4345 2.3312
No log 0.4583 22 4.8446 0.0222 4.8446 2.2011
No log 0.5 24 2.9379 0.1071 2.9379 1.7140
No log 0.5417 26 2.0248 0.2361 2.0248 1.4230
No log 0.5833 28 1.7140 0.3252 1.7140 1.3092
No log 0.625 30 1.9322 0.3594 1.9322 1.3900
No log 0.6667 32 2.1066 0.2794 2.1066 1.4514
No log 0.7083 34 2.2637 0.1944 2.2637 1.5046
No log 0.75 36 2.5106 0.1711 2.5106 1.5845
No log 0.7917 38 2.5837 0.1688 2.5837 1.6074
No log 0.8333 40 2.3657 0.2051 2.3657 1.5381
No log 0.875 42 2.0623 0.2933 2.0623 1.4361
No log 0.9167 44 1.9258 0.3816 1.9258 1.3877
No log 0.9583 46 1.5098 0.4173 1.5098 1.2287
No log 1.0 48 1.3606 0.5036 1.3606 1.1664
No log 1.0417 50 1.4413 0.4348 1.4413 1.2005
No log 1.0833 52 1.8443 0.3733 1.8443 1.3581
No log 1.125 54 2.5324 0.2963 2.5324 1.5914
No log 1.1667 56 2.8803 0.2857 2.8803 1.6972
No log 1.2083 58 3.1933 0.2620 3.1933 1.7870
No log 1.25 60 2.7820 0.2885 2.7820 1.6679
No log 1.2917 62 1.9885 0.3766 1.9885 1.4102
No log 1.3333 64 1.5583 0.4148 1.5583 1.2483
No log 1.375 66 1.5171 0.4571 1.5171 1.2317
No log 1.4167 68 1.5930 0.4615 1.5930 1.2622
No log 1.4583 70 2.0784 0.4138 2.0784 1.4417
No log 1.5 72 2.1782 0.4045 2.1782 1.4759
No log 1.5417 74 2.0859 0.3886 2.0859 1.4442
No log 1.5833 76 1.8005 0.4337 1.8005 1.3418
No log 1.625 78 1.7811 0.4586 1.7811 1.3346
No log 1.6667 80 2.0336 0.3523 2.0336 1.4261
No log 1.7083 82 2.5223 0.2963 2.5223 1.5882
No log 1.75 84 2.5739 0.2917 2.5739 1.6043
No log 1.7917 86 2.1908 0.3102 2.1908 1.4801
No log 1.8333 88 1.5318 0.4714 1.5318 1.2377
No log 1.875 90 1.3239 0.4741 1.3239 1.1506
No log 1.9167 92 1.2800 0.4478 1.2800 1.1314
No log 1.9583 94 1.3641 0.4895 1.3641 1.1680
No log 2.0 96 1.6294 0.4444 1.6294 1.2765
No log 2.0417 98 1.8458 0.3816 1.8458 1.3586
No log 2.0833 100 1.8221 0.3949 1.8221 1.3498
No log 2.125 102 1.9362 0.4049 1.9362 1.3915
No log 2.1667 104 2.3522 0.4000 2.3522 1.5337
No log 2.2083 106 2.4160 0.3673 2.4160 1.5544
No log 2.25 108 1.9185 0.4457 1.9185 1.3851
No log 2.2917 110 1.9008 0.4157 1.9008 1.3787
No log 2.3333 112 2.1559 0.4022 2.1559 1.4683
No log 2.375 114 3.0548 0.2844 3.0548 1.7478
No log 2.4167 116 2.6484 0.3779 2.6484 1.6274
No log 2.4583 118 1.5566 0.5424 1.5566 1.2476
No log 2.5 120 0.9320 0.6294 0.9320 0.9654
No log 2.5417 122 0.8883 0.6853 0.8883 0.9425
No log 2.5833 124 0.8323 0.6993 0.8323 0.9123
No log 2.625 126 0.8833 0.6345 0.8833 0.9398
No log 2.6667 128 0.9005 0.6621 0.9005 0.9490
No log 2.7083 130 0.8994 0.6438 0.8994 0.9484
No log 2.75 132 1.0196 0.6013 1.0196 1.0097
No log 2.7917 134 1.0943 0.6038 1.0943 1.0461
No log 2.8333 136 1.0999 0.6380 1.0999 1.0488
No log 2.875 138 1.3409 0.6118 1.3409 1.1580
No log 2.9167 140 1.3554 0.6369 1.3554 1.1642
No log 2.9583 142 1.4411 0.6180 1.4411 1.2005
No log 3.0 144 1.1088 0.6941 1.1088 1.0530
No log 3.0417 146 0.8444 0.6351 0.8444 0.9189
No log 3.0833 148 0.8406 0.6667 0.8406 0.9168
No log 3.125 150 0.8657 0.6423 0.8657 0.9304
No log 3.1667 152 0.9031 0.6901 0.9031 0.9503
No log 3.2083 154 1.3122 0.5749 1.3122 1.1455
No log 3.25 156 1.5197 0.5412 1.5197 1.2328
No log 3.2917 158 1.3430 0.5409 1.3430 1.1589
No log 3.3333 160 0.9828 0.6418 0.9828 0.9914
No log 3.375 162 0.8741 0.6212 0.8741 0.9349
No log 3.4167 164 0.8762 0.6412 0.8762 0.9360
No log 3.4583 166 0.9025 0.6412 0.9025 0.9500
No log 3.5 168 0.9026 0.6165 0.9026 0.9500
No log 3.5417 170 0.9130 0.6061 0.9130 0.9555
No log 3.5833 172 0.9066 0.6212 0.9066 0.9521
No log 3.625 174 1.0116 0.625 1.0116 1.0058
No log 3.6667 176 1.2988 0.6180 1.2988 1.1396
No log 3.7083 178 1.4060 0.6304 1.4060 1.1857
No log 3.75 180 1.5689 0.4975 1.5689 1.2525
No log 3.7917 182 1.5038 0.5668 1.5038 1.2263
No log 3.8333 184 1.5005 0.6087 1.5005 1.2250
No log 3.875 186 1.2419 0.6477 1.2419 1.1144
No log 3.9167 188 1.2367 0.6480 1.2367 1.1121
No log 3.9583 190 1.3361 0.6559 1.3361 1.1559
No log 4.0 192 1.2104 0.6480 1.2104 1.1002
No log 4.0417 194 1.2484 0.6519 1.2484 1.1173
No log 4.0833 196 1.1251 0.6506 1.1251 1.0607
No log 4.125 198 1.1290 0.6369 1.1290 1.0626
No log 4.1667 200 1.3516 0.5399 1.3516 1.1626
No log 4.2083 202 1.4981 0.5140 1.4981 1.2240
No log 4.25 204 1.5046 0.5 1.5046 1.2266
No log 4.2917 206 1.2477 0.5844 1.2477 1.1170
No log 4.3333 208 1.0948 0.5468 1.0948 1.0463
No log 4.375 210 1.0547 0.5414 1.0547 1.0270
No log 4.4167 212 1.0475 0.5588 1.0475 1.0235
No log 4.4583 214 1.0598 0.6 1.0598 1.0295
No log 4.5 216 1.0705 0.6301 1.0705 1.0347
No log 4.5417 218 1.2135 0.6420 1.2135 1.1016
No log 4.5833 220 1.0834 0.6710 1.0834 1.0409
No log 4.625 222 0.9543 0.7020 0.9543 0.9769
No log 4.6667 224 0.9281 0.6207 0.9281 0.9634
No log 4.7083 226 0.8275 0.6806 0.8275 0.9097
No log 4.75 228 0.8891 0.6412 0.8891 0.9429
No log 4.7917 230 1.1046 0.5846 1.1046 1.0510
No log 4.8333 232 1.1428 0.5846 1.1428 1.0690
No log 4.875 234 0.9934 0.6269 0.9934 0.9967
No log 4.9167 236 0.8512 0.7 0.8512 0.9226
No log 4.9583 238 0.8024 0.7042 0.8024 0.8958
No log 5.0 240 0.7716 0.7347 0.7716 0.8784
No log 5.0417 242 0.7813 0.7297 0.7813 0.8839
No log 5.0833 244 0.7703 0.7347 0.7703 0.8776
No log 5.125 246 0.7813 0.7050 0.7813 0.8839
No log 5.1667 248 0.9953 0.5735 0.9953 0.9976
No log 5.2083 250 1.1336 0.5185 1.1336 1.0647
No log 5.25 252 1.0596 0.5606 1.0596 1.0293
No log 5.2917 254 0.9009 0.6718 0.9009 0.9492
No log 5.3333 256 0.8228 0.6716 0.8228 0.9071
No log 5.375 258 0.8279 0.6715 0.8279 0.9099
No log 5.4167 260 0.8522 0.6412 0.8522 0.9231
No log 5.4583 262 0.9017 0.625 0.9017 0.9496
No log 5.5 264 0.9062 0.625 0.9062 0.9519
No log 5.5417 266 0.9187 0.6622 0.9187 0.9585
No log 5.5833 268 0.9397 0.6351 0.9397 0.9694
No log 5.625 270 0.8968 0.6294 0.8968 0.9470
No log 5.6667 272 0.8629 0.7133 0.8629 0.9289
No log 5.7083 274 0.8710 0.6839 0.8710 0.9333
No log 5.75 276 0.9610 0.7059 0.9610 0.9803
No log 5.7917 278 0.9745 0.6857 0.9745 0.9872
No log 5.8333 280 1.0197 0.6857 1.0197 1.0098
No log 5.875 282 0.9568 0.7011 0.9568 0.9781
No log 5.9167 284 0.9159 0.7135 0.9159 0.9570
No log 5.9583 286 0.8465 0.7229 0.8465 0.9200
No log 6.0 288 0.8156 0.7044 0.8156 0.9031
No log 6.0417 290 0.8399 0.7394 0.8399 0.9165
No log 6.0833 292 0.8636 0.7337 0.8636 0.9293
No log 6.125 294 0.9360 0.6824 0.9360 0.9675
No log 6.1667 296 0.8684 0.7296 0.8684 0.9319
No log 6.2083 298 0.7866 0.7763 0.7866 0.8869
No log 6.25 300 0.8167 0.6857 0.8167 0.9037
No log 6.2917 302 0.8424 0.6715 0.8424 0.9178
No log 6.3333 304 0.8129 0.6763 0.8129 0.9016
No log 6.375 306 0.8359 0.7125 0.8359 0.9143
No log 6.4167 308 0.8996 0.7081 0.8996 0.9485
No log 6.4583 310 0.9110 0.7160 0.9110 0.9544
No log 6.5 312 0.9963 0.7262 0.9963 0.9982
No log 6.5417 314 1.0671 0.7126 1.0671 1.0330
No log 6.5833 316 1.0547 0.7044 1.0547 1.0270
No log 6.625 318 1.0859 0.6748 1.0859 1.0421
No log 6.6667 320 0.9561 0.6624 0.9561 0.9778
No log 6.7083 322 0.8200 0.6765 0.8200 0.9056
No log 6.75 324 0.7964 0.7059 0.7964 0.8924
No log 6.7917 326 0.8895 0.6667 0.8895 0.9431
No log 6.8333 328 0.8948 0.6618 0.8948 0.9459
No log 6.875 330 0.7941 0.7361 0.7941 0.8911
No log 6.9167 332 0.7944 0.7453 0.7944 0.8913
No log 6.9583 334 0.9602 0.7263 0.9602 0.9799
No log 7.0 336 1.1501 0.6495 1.1501 1.0724
No log 7.0417 338 1.0743 0.7041 1.0743 1.0365
No log 7.0833 340 0.8419 0.7684 0.8419 0.9175
No log 7.125 342 0.7410 0.7436 0.7410 0.8608
No log 7.1667 344 0.7910 0.7436 0.7910 0.8894
No log 7.2083 346 0.8386 0.7105 0.8386 0.9157
No log 7.25 348 0.8423 0.7333 0.8423 0.9178
No log 7.2917 350 0.8417 0.7248 0.8417 0.9174
No log 7.3333 352 0.8150 0.7190 0.8150 0.9028
No log 7.375 354 0.8065 0.7421 0.8065 0.8981
No log 7.4167 356 0.8129 0.7442 0.8129 0.9016
No log 7.4583 358 0.9023 0.7356 0.9023 0.9499
No log 7.5 360 1.0471 0.7083 1.0471 1.0233
No log 7.5417 362 1.1344 0.6804 1.1344 1.0651
No log 7.5833 364 1.0833 0.6842 1.0833 1.0408
No log 7.625 366 0.9004 0.6941 0.9004 0.9489
No log 7.6667 368 0.7699 0.7342 0.7699 0.8775
No log 7.7083 370 0.7367 0.7 0.7367 0.8583
No log 7.75 372 0.8430 0.6618 0.8430 0.9182
No log 7.7917 374 0.9298 0.6429 0.9298 0.9643
No log 7.8333 376 0.9124 0.6429 0.9124 0.9552
No log 7.875 378 0.8144 0.6812 0.8144 0.9024
No log 7.9167 380 0.7282 0.7050 0.7282 0.8533
No log 7.9583 382 0.6961 0.7619 0.6961 0.8343
No log 8.0 384 0.7337 0.7558 0.7337 0.8566
No log 8.0417 386 0.7395 0.7543 0.7395 0.8599
No log 8.0833 388 0.7298 0.7816 0.7298 0.8543
No log 8.125 390 0.7267 0.7738 0.7267 0.8525
No log 8.1667 392 0.7427 0.7632 0.7427 0.8618
No log 8.2083 394 0.7346 0.7183 0.7346 0.8571
No log 8.25 396 0.7495 0.7050 0.7495 0.8658
No log 8.2917 398 0.7361 0.7050 0.7361 0.8580
No log 8.3333 400 0.7350 0.6963 0.7350 0.8573
No log 8.375 402 0.6992 0.7164 0.6992 0.8362
No log 8.4167 404 0.6741 0.7299 0.6741 0.8210
No log 8.4583 406 0.6751 0.7353 0.6751 0.8216
No log 8.5 408 0.7346 0.7068 0.7346 0.8571
No log 8.5417 410 0.7620 0.6866 0.7620 0.8729
No log 8.5833 412 0.7441 0.6818 0.7441 0.8626
No log 8.625 414 0.7568 0.7015 0.7568 0.8699
No log 8.6667 416 0.7737 0.7153 0.7737 0.8796
No log 8.7083 418 0.8367 0.6565 0.8367 0.9147
No log 8.75 420 0.9247 0.6364 0.9247 0.9616
No log 8.7917 422 0.9518 0.6429 0.9518 0.9756
No log 8.8333 424 0.8492 0.7211 0.8492 0.9215
No log 8.875 426 0.7618 0.7324 0.7618 0.8728
No log 8.9167 428 0.7298 0.7465 0.7298 0.8543
No log 8.9583 430 0.7269 0.6619 0.7269 0.8526
No log 9.0 432 0.7169 0.6950 0.7169 0.8467
No log 9.0417 434 0.6643 0.7660 0.6643 0.8150
No log 9.0833 436 0.6543 0.7407 0.6543 0.8089
No log 9.125 438 0.6587 0.7313 0.6587 0.8116
No log 9.1667 440 0.6537 0.7733 0.6537 0.8085
No log 9.2083 442 0.6622 0.7815 0.6622 0.8138
No log 9.25 444 0.7143 0.7722 0.7143 0.8451
No log 9.2917 446 0.8064 0.7586 0.8064 0.8980
No log 9.3333 448 0.8613 0.6936 0.8613 0.9281
No log 9.375 450 0.8930 0.6860 0.8930 0.9450
No log 9.4167 452 0.8931 0.7093 0.8931 0.9450
No log 9.4583 454 0.8528 0.7226 0.8528 0.9235
No log 9.5 456 0.7795 0.7297 0.7795 0.8829
No log 9.5417 458 0.7423 0.7092 0.7423 0.8616
No log 9.5833 460 0.7480 0.7023 0.7480 0.8649
No log 9.625 462 0.7644 0.7077 0.7644 0.8743
No log 9.6667 464 0.8234 0.7023 0.8234 0.9074
No log 9.7083 466 0.9041 0.6515 0.9041 0.9508
No log 9.75 468 0.9526 0.6515 0.9526 0.9760
No log 9.7917 470 0.8744 0.6718 0.8744 0.9351
No log 9.8333 472 0.7790 0.6917 0.7790 0.8826
No log 9.875 474 0.8336 0.7097 0.8336 0.9130
No log 9.9167 476 1.0839 0.6235 1.0839 1.0411
No log 9.9583 478 1.1330 0.6480 1.1330 1.0644
No log 10.0 480 1.0124 0.7086 1.0124 1.0062
No log 10.0417 482 0.8732 0.7126 0.8732 0.9345
No log 10.0833 484 0.8387 0.7251 0.8387 0.9158
No log 10.125 486 0.9397 0.72 0.9397 0.9694
No log 10.1667 488 1.2520 0.6556 1.2520 1.1189
No log 10.2083 490 1.3704 0.5914 1.3704 1.1706
No log 10.25 492 1.2877 0.6344 1.2877 1.1348
No log 10.2917 494 1.1275 0.6593 1.1275 1.0618
No log 10.3333 496 0.8921 0.7314 0.8921 0.9445
No log 10.375 498 0.7064 0.75 0.7064 0.8405
0.4401 10.4167 500 0.7085 0.7483 0.7085 0.8417
0.4401 10.4583 502 0.7211 0.7194 0.7211 0.8492
0.4401 10.5 504 0.6778 0.7808 0.6778 0.8233
0.4401 10.5417 506 0.6685 0.8 0.6685 0.8176
0.4401 10.5833 508 0.6793 0.7947 0.6793 0.8242
0.4401 10.625 510 0.7207 0.7397 0.7207 0.8490
0.4401 10.6667 512 0.8224 0.6619 0.8224 0.9069
0.4401 10.7083 514 0.8313 0.6471 0.8313 0.9117
0.4401 10.75 516 0.7440 0.7133 0.7440 0.8625
0.4401 10.7917 518 0.6958 0.7639 0.6958 0.8341
0.4401 10.8333 520 0.6882 0.7703 0.6882 0.8296
0.4401 10.875 522 0.6940 0.7703 0.6940 0.8331
0.4401 10.9167 524 0.6921 0.7815 0.6921 0.8319
0.4401 10.9583 526 0.6906 0.7815 0.6906 0.8311
0.4401 11.0 528 0.6992 0.7947 0.6992 0.8362
0.4401 11.0417 530 0.7038 0.7974 0.7038 0.8389
0.4401 11.0833 532 0.7378 0.7643 0.7378 0.8590
0.4401 11.125 534 0.7752 0.7436 0.7752 0.8804
0.4401 11.1667 536 0.7761 0.7324 0.7761 0.8809
0.4401 11.2083 538 0.7567 0.7111 0.7567 0.8699
0.4401 11.25 540 0.7446 0.7259 0.7446 0.8629
0.4401 11.2917 542 0.7442 0.7259 0.7442 0.8627

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k10_task1_organization

Finetuned
(4019)
this model