ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k17_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9345
  • Qwk: 0.6479
  • Mse: 0.9345
  • Rmse: 0.9667

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.025 2 6.9243 0.0057 6.9243 2.6314
No log 0.05 4 4.4589 0.0598 4.4589 2.1116
No log 0.075 6 3.2248 0.0339 3.2248 1.7958
No log 0.1 8 3.4562 0.0109 3.4562 1.8591
No log 0.125 10 2.9791 0.0355 2.9791 1.7260
No log 0.15 12 2.1223 0.1613 2.1223 1.4568
No log 0.175 14 1.7803 0.1852 1.7803 1.3343
No log 0.2 16 1.6969 0.1524 1.6969 1.3027
No log 0.225 18 1.7249 0.0935 1.7249 1.3133
No log 0.25 20 1.6778 0.1509 1.6778 1.2953
No log 0.275 22 2.0599 0.1832 2.0599 1.4352
No log 0.3 24 2.7892 0.0637 2.7892 1.6701
No log 0.325 26 3.1103 0.0723 3.1103 1.7636
No log 0.35 28 2.1807 0.2254 2.1807 1.4767
No log 0.375 30 1.6242 0.352 1.6242 1.2745
No log 0.4 32 1.7380 0.4 1.7380 1.3183
No log 0.425 34 2.0077 0.2963 2.0077 1.4169
No log 0.45 36 2.0376 0.2647 2.0376 1.4274
No log 0.475 38 2.0028 0.2647 2.0028 1.4152
No log 0.5 40 2.0183 0.2628 2.0183 1.4207
No log 0.525 42 2.1133 0.2394 2.1133 1.4537
No log 0.55 44 2.3019 0.3000 2.3019 1.5172
No log 0.575 46 1.9983 0.3537 1.9983 1.4136
No log 0.6 48 1.9021 0.3000 1.9021 1.3792
No log 0.625 50 1.8666 0.2815 1.8666 1.3662
No log 0.65 52 1.7428 0.2754 1.7428 1.3201
No log 0.675 54 1.7835 0.4151 1.7835 1.3355
No log 0.7 56 2.1129 0.4186 2.1129 1.4536
No log 0.725 58 1.9307 0.375 1.9307 1.3895
No log 0.75 60 2.0848 0.3659 2.0848 1.4439
No log 0.775 62 2.3456 0.3103 2.3456 1.5315
No log 0.8 64 2.2680 0.3390 2.2680 1.5060
No log 0.825 66 2.0697 0.3550 2.0697 1.4387
No log 0.85 68 1.7421 0.4545 1.7421 1.3199
No log 0.875 70 1.3780 0.5 1.3780 1.1739
No log 0.9 72 1.1799 0.4567 1.1799 1.0862
No log 0.925 74 1.1683 0.5038 1.1683 1.0809
No log 0.95 76 1.5273 0.4713 1.5273 1.2358
No log 0.975 78 2.0899 0.3616 2.0899 1.4457
No log 1.0 80 2.2924 0.3316 2.2924 1.5141
No log 1.025 82 2.2035 0.3503 2.2035 1.4844
No log 1.05 84 1.9050 0.4 1.9050 1.3802
No log 1.075 86 1.5546 0.4429 1.5546 1.2468
No log 1.1 88 1.4453 0.4526 1.4453 1.2022
No log 1.125 90 1.4482 0.4143 1.4482 1.2034
No log 1.15 92 1.3834 0.4173 1.3834 1.1762
No log 1.175 94 1.1441 0.5455 1.1441 1.0696
No log 1.2 96 1.0789 0.5303 1.0789 1.0387
No log 1.225 98 1.2126 0.5152 1.2126 1.1012
No log 1.25 100 1.4831 0.4605 1.4831 1.2178
No log 1.275 102 1.8353 0.4048 1.8353 1.3547
No log 1.3 104 1.9087 0.4 1.9087 1.3816
No log 1.325 106 1.5602 0.4198 1.5602 1.2491
No log 1.35 108 1.3721 0.5065 1.3721 1.1713
No log 1.375 110 1.3700 0.5375 1.3700 1.1705
No log 1.4 112 1.4249 0.5366 1.4249 1.1937
No log 1.425 114 1.2329 0.6012 1.2329 1.1103
No log 1.45 116 1.0556 0.6364 1.0556 1.0274
No log 1.475 118 1.0616 0.6364 1.0616 1.0303
No log 1.5 120 0.9382 0.6405 0.9382 0.9686
No log 1.525 122 0.7752 0.7432 0.7752 0.8804
No log 1.55 124 0.7700 0.75 0.7700 0.8775
No log 1.575 126 0.8056 0.7179 0.8056 0.8975
No log 1.6 128 1.0156 0.6786 1.0156 1.0078
No log 1.625 130 1.4812 0.5393 1.4812 1.2171
No log 1.65 132 1.5141 0.5311 1.5141 1.2305
No log 1.675 134 1.0088 0.6579 1.0088 1.0044
No log 1.7 136 0.7418 0.6567 0.7418 0.8613
No log 1.725 138 0.8206 0.7164 0.8206 0.9059
No log 1.75 140 0.8650 0.6970 0.8650 0.9300
No log 1.775 142 0.7714 0.7068 0.7714 0.8783
No log 1.8 144 0.7046 0.7586 0.7046 0.8394
No log 1.825 146 0.9499 0.6667 0.9499 0.9746
No log 1.85 148 1.0581 0.6667 1.0581 1.0286
No log 1.875 150 0.8550 0.6918 0.8550 0.9247
No log 1.9 152 0.7124 0.7582 0.7124 0.8440
No log 1.925 154 0.6847 0.7682 0.6847 0.8274
No log 1.95 156 0.7596 0.7285 0.7596 0.8716
No log 1.975 158 0.7798 0.7067 0.7798 0.8831
No log 2.0 160 0.7041 0.7432 0.7041 0.8391
No log 2.025 162 0.7458 0.7206 0.7458 0.8636
No log 2.05 164 0.8423 0.6809 0.8423 0.9178
No log 2.075 166 0.7973 0.7183 0.7973 0.8929
No log 2.1 168 0.7501 0.7383 0.7501 0.8661
No log 2.125 170 0.8552 0.6585 0.8552 0.9248
No log 2.15 172 0.9710 0.6746 0.9710 0.9854
No log 2.175 174 0.8808 0.6826 0.8808 0.9385
No log 2.2 176 0.7785 0.6714 0.7785 0.8823
No log 2.225 178 0.7963 0.6715 0.7963 0.8924
No log 2.25 180 0.8131 0.6715 0.8131 0.9017
No log 2.275 182 0.8412 0.6475 0.8412 0.9171
No log 2.3 184 0.9305 0.6351 0.9305 0.9646
No log 2.325 186 1.0409 0.7044 1.0409 1.0202
No log 2.35 188 1.0162 0.6711 1.0162 1.0081
No log 2.375 190 0.9737 0.6043 0.9737 0.9868
No log 2.4 192 1.0035 0.6131 1.0035 1.0017
No log 2.425 194 1.0219 0.6364 1.0219 1.0109
No log 2.45 196 0.9802 0.6357 0.9802 0.9900
No log 2.475 198 0.9225 0.6131 0.9225 0.9605
No log 2.5 200 1.0004 0.6795 1.0004 1.0002
No log 2.525 202 1.1716 0.6628 1.1716 1.0824
No log 2.55 204 1.1945 0.6480 1.1945 1.0929
No log 2.575 206 1.1492 0.6851 1.1492 1.0720
No log 2.6 208 1.0793 0.6936 1.0793 1.0389
No log 2.625 210 1.0018 0.7024 1.0018 1.0009
No log 2.65 212 0.9256 0.6753 0.9256 0.9621
No log 2.675 214 0.9281 0.6452 0.9281 0.9634
No log 2.7 216 0.9090 0.6494 0.9090 0.9534
No log 2.725 218 0.9468 0.6452 0.9468 0.9730
No log 2.75 220 0.9888 0.6415 0.9888 0.9944
No log 2.775 222 1.0248 0.6625 1.0248 1.0123
No log 2.8 224 1.2536 0.6391 1.2536 1.1197
No log 2.825 226 1.2255 0.6386 1.2255 1.1070
No log 2.85 228 1.0643 0.6303 1.0643 1.0317
No log 2.875 230 0.9693 0.6624 0.9693 0.9845
No log 2.9 232 0.9937 0.675 0.9937 0.9969
No log 2.925 234 0.9816 0.6832 0.9816 0.9908
No log 2.95 236 0.8547 0.6710 0.8547 0.9245
No log 2.975 238 0.7671 0.6755 0.7671 0.8758
No log 3.0 240 0.7587 0.6857 0.7587 0.8710
No log 3.025 242 0.8675 0.6423 0.8675 0.9314
No log 3.05 244 0.9594 0.6165 0.9594 0.9795
No log 3.075 246 0.9814 0.6418 0.9814 0.9906
No log 3.1 248 0.8420 0.6906 0.8420 0.9176
No log 3.125 250 0.7147 0.7211 0.7147 0.8454
No log 3.15 252 0.7571 0.7179 0.7571 0.8701
No log 3.175 254 0.8422 0.7030 0.8422 0.9177
No log 3.2 256 0.7700 0.7105 0.7700 0.8775
No log 3.225 258 0.7631 0.7143 0.7631 0.8735
No log 3.25 260 0.6845 0.7105 0.6845 0.8273
No log 3.275 262 0.6437 0.7552 0.6437 0.8023
No log 3.3 264 0.7262 0.7376 0.7262 0.8522
No log 3.325 266 0.7578 0.7273 0.7578 0.8705
No log 3.35 268 0.7396 0.7361 0.7396 0.8600
No log 3.375 270 0.7366 0.7361 0.7366 0.8583
No log 3.4 272 0.7089 0.7639 0.7089 0.8420
No log 3.425 274 0.7258 0.7682 0.7258 0.8519
No log 3.45 276 0.7400 0.7375 0.7400 0.8602
No log 3.475 278 0.7730 0.7394 0.7730 0.8792
No log 3.5 280 0.7920 0.7176 0.7920 0.8900
No log 3.525 282 0.7953 0.7356 0.7953 0.8918
No log 3.55 284 0.7616 0.7262 0.7616 0.8727
No log 3.575 286 0.7649 0.7170 0.7649 0.8746
No log 3.6 288 0.8849 0.7205 0.8849 0.9407
No log 3.625 290 1.2687 0.6257 1.2687 1.1264
No log 3.65 292 1.4718 0.5604 1.4718 1.2132
No log 3.675 294 1.3297 0.6108 1.3297 1.1531
No log 3.7 296 0.9834 0.6197 0.9834 0.9916
No log 3.725 298 0.7916 0.6571 0.7916 0.8897
No log 3.75 300 0.7554 0.6957 0.7554 0.8691
No log 3.775 302 0.7698 0.7059 0.7698 0.8774
No log 3.8 304 0.7660 0.7361 0.7660 0.8752
No log 3.825 306 0.7704 0.7162 0.7704 0.8777
No log 3.85 308 0.7573 0.7692 0.7573 0.8702
No log 3.875 310 0.6934 0.7651 0.6934 0.8327
No log 3.9 312 0.7005 0.7361 0.7005 0.8370
No log 3.925 314 0.7208 0.7552 0.7208 0.8490
No log 3.95 316 0.7082 0.7448 0.7082 0.8415
No log 3.975 318 0.7287 0.7785 0.7287 0.8537
No log 4.0 320 0.7816 0.7297 0.7816 0.8841
No log 4.025 322 0.7346 0.7483 0.7346 0.8571
No log 4.05 324 0.6911 0.7733 0.6911 0.8313
No log 4.075 326 0.6845 0.7632 0.6845 0.8273
No log 4.1 328 0.6983 0.75 0.6983 0.8356
No log 4.125 330 0.7217 0.7517 0.7217 0.8495
No log 4.15 332 0.7955 0.7347 0.7955 0.8919
No log 4.175 334 0.8853 0.7123 0.8853 0.9409
No log 4.2 336 0.8723 0.6944 0.8723 0.9340
No log 4.225 338 0.7756 0.7532 0.7756 0.8807
No log 4.25 340 0.7569 0.7886 0.7569 0.8700
No log 4.275 342 0.7640 0.7912 0.7640 0.8741
No log 4.3 344 0.6772 0.8222 0.6772 0.8229
No log 4.325 346 0.6137 0.8023 0.6137 0.7834
No log 4.35 348 0.6192 0.8090 0.6192 0.7869
No log 4.375 350 0.6922 0.7956 0.6922 0.8320
No log 4.4 352 0.6823 0.8156 0.6823 0.8260
No log 4.425 354 0.6567 0.8182 0.6567 0.8104
No log 4.45 356 0.6127 0.8161 0.6127 0.7828
No log 4.475 358 0.6170 0.8024 0.6170 0.7855
No log 4.5 360 0.6777 0.7927 0.6777 0.8232
No log 4.525 362 0.6912 0.7843 0.6912 0.8314
No log 4.55 364 0.7638 0.7342 0.7638 0.8740
No log 4.575 366 0.7442 0.7547 0.7442 0.8627
No log 4.6 368 0.6511 0.7651 0.6511 0.8069
No log 4.625 370 0.6586 0.7552 0.6586 0.8116
No log 4.65 372 0.6695 0.7448 0.6695 0.8182
No log 4.675 374 0.6797 0.75 0.6797 0.8244
No log 4.7 376 0.6829 0.7361 0.6829 0.8264
No log 4.725 378 0.6898 0.76 0.6898 0.8305
No log 4.75 380 0.7787 0.7595 0.7787 0.8824
No log 4.775 382 0.8592 0.6707 0.8592 0.9269
No log 4.8 384 0.8016 0.6962 0.8016 0.8953
No log 4.825 386 0.7070 0.7671 0.7070 0.8409
No log 4.85 388 0.7321 0.7234 0.7321 0.8556
No log 4.875 390 0.7748 0.6912 0.7748 0.8802
No log 4.9 392 0.8278 0.6912 0.8278 0.9099
No log 4.925 394 0.8615 0.6861 0.8615 0.9281
No log 4.95 396 0.8161 0.6912 0.8161 0.9034
No log 4.975 398 0.7609 0.7338 0.7609 0.8723
No log 5.0 400 0.7234 0.7536 0.7234 0.8505
No log 5.025 402 0.7392 0.7338 0.7392 0.8598
No log 5.05 404 0.7202 0.7413 0.7202 0.8486
No log 5.075 406 0.7350 0.7376 0.7350 0.8574
No log 5.1 408 0.7668 0.7111 0.7668 0.8757
No log 5.125 410 0.7573 0.7286 0.7573 0.8702
No log 5.15 412 0.7756 0.7059 0.7756 0.8807
No log 5.175 414 0.8083 0.7015 0.8083 0.8991
No log 5.2 416 0.8552 0.6765 0.8552 0.9248
No log 5.225 418 0.7926 0.6963 0.7926 0.8903
No log 5.25 420 0.7626 0.7299 0.7626 0.8733
No log 5.275 422 0.7676 0.7194 0.7676 0.8761
No log 5.3 424 0.7776 0.7206 0.7776 0.8818
No log 5.325 426 0.7738 0.7068 0.7738 0.8797
No log 5.35 428 0.7783 0.6870 0.7783 0.8822
No log 5.375 430 0.7475 0.7361 0.7475 0.8646
No log 5.4 432 0.6934 0.7534 0.6934 0.8327
No log 5.425 434 0.6998 0.7712 0.6998 0.8366
No log 5.45 436 0.7016 0.7417 0.7016 0.8376
No log 5.475 438 0.6716 0.7733 0.6716 0.8195
No log 5.5 440 0.6676 0.7310 0.6676 0.8170
No log 5.525 442 0.7133 0.7183 0.7133 0.8446
No log 5.55 444 0.7204 0.7050 0.7204 0.8488
No log 5.575 446 0.6965 0.7376 0.6965 0.8345
No log 5.6 448 0.7079 0.7338 0.7079 0.8414
No log 5.625 450 0.7300 0.7338 0.7300 0.8544
No log 5.65 452 0.7130 0.7536 0.7130 0.8444
No log 5.675 454 0.6943 0.7536 0.6943 0.8332
No log 5.7 456 0.6892 0.7536 0.6892 0.8302
No log 5.725 458 0.6858 0.7338 0.6858 0.8281
No log 5.75 460 0.7051 0.7123 0.7051 0.8397
No log 5.775 462 0.6759 0.7451 0.6759 0.8222
No log 5.8 464 0.6188 0.7755 0.6188 0.7866
No log 5.825 466 0.6087 0.75 0.6087 0.7802
No log 5.85 468 0.6925 0.7050 0.6925 0.8322
No log 5.875 470 0.7310 0.7050 0.7310 0.8550
No log 5.9 472 0.6913 0.7183 0.6913 0.8315
No log 5.925 474 0.6265 0.7671 0.6265 0.7915
No log 5.95 476 0.6170 0.7898 0.6170 0.7855
No log 5.975 478 0.6368 0.7821 0.6368 0.7980
No log 6.0 480 0.6507 0.7947 0.6507 0.8067
No log 6.025 482 0.7133 0.7552 0.7133 0.8446
No log 6.05 484 0.8191 0.6815 0.8191 0.9050
No log 6.075 486 0.8914 0.6815 0.8914 0.9442
No log 6.1 488 0.8558 0.6767 0.8558 0.9251
No log 6.125 490 0.7923 0.7068 0.7923 0.8901
No log 6.15 492 0.7787 0.6615 0.7787 0.8824
No log 6.175 494 0.7851 0.6667 0.7851 0.8860
No log 6.2 496 0.7775 0.6418 0.7775 0.8818
No log 6.225 498 0.7473 0.6412 0.7473 0.8645
0.4333 6.25 500 0.7187 0.7164 0.7187 0.8477
0.4333 6.275 502 0.7317 0.6912 0.7317 0.8554
0.4333 6.3 504 0.6902 0.7050 0.6902 0.8308
0.4333 6.325 506 0.6182 0.7260 0.6182 0.7863
0.4333 6.35 508 0.6023 0.7703 0.6023 0.7761
0.4333 6.375 510 0.6700 0.7742 0.6700 0.8185
0.4333 6.4 512 0.7286 0.7389 0.7286 0.8536
0.4333 6.425 514 0.7161 0.7468 0.7161 0.8462
0.4333 6.45 516 0.6531 0.7821 0.6531 0.8081
0.4333 6.475 518 0.6041 0.8025 0.6041 0.7772
0.4333 6.5 520 0.5964 0.8101 0.5964 0.7723
0.4333 6.525 522 0.6136 0.7547 0.6136 0.7833
0.4333 6.55 524 0.6906 0.7778 0.6906 0.8310
0.4333 6.575 526 0.8741 0.7135 0.8741 0.9349
0.4333 6.6 528 0.9151 0.7176 0.9151 0.9566
0.4333 6.625 530 0.8106 0.6792 0.8106 0.9004
0.4333 6.65 532 0.6681 0.7895 0.6681 0.8174
0.4333 6.675 534 0.6841 0.7338 0.6841 0.8271
0.4333 6.7 536 0.7486 0.7059 0.7486 0.8652
0.4333 6.725 538 0.7846 0.7313 0.7846 0.8858
0.4333 6.75 540 0.8117 0.6667 0.8117 0.9010
0.4333 6.775 542 0.8347 0.6515 0.8347 0.9136
0.4333 6.8 544 0.8728 0.5802 0.8728 0.9342
0.4333 6.825 546 0.8911 0.6074 0.8911 0.9440
0.4333 6.85 548 0.9345 0.6479 0.9345 0.9667

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k17_task1_organization

Finetuned
(4023)
this model