ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k14_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8269
  • Qwk: 0.7183
  • Mse: 0.8269
  • Rmse: 0.9094

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0303 2 9.5179 0.0039 9.5179 3.0851
No log 0.0606 4 5.9262 0.0 5.9262 2.4344
No log 0.0909 6 3.9079 0.0625 3.9079 1.9768
No log 0.1212 8 4.3761 -0.0179 4.3761 2.0919
No log 0.1515 10 3.8536 -0.0535 3.8536 1.9631
No log 0.1818 12 2.0382 0.1333 2.0382 1.4277
No log 0.2121 14 1.7004 0.1682 1.7004 1.3040
No log 0.2424 16 1.8428 0.2698 1.8428 1.3575
No log 0.2727 18 2.1684 0.1931 2.1684 1.4726
No log 0.3030 20 2.3120 0.1216 2.3120 1.5205
No log 0.3333 22 2.0868 0.2676 2.0868 1.4446
No log 0.3636 24 1.7943 0.3333 1.7943 1.3395
No log 0.3939 26 1.4391 0.3607 1.4391 1.1996
No log 0.4242 28 1.3460 0.3826 1.3460 1.1602
No log 0.4545 30 1.1945 0.4483 1.1945 1.0929
No log 0.4848 32 1.3009 0.3091 1.3009 1.1406
No log 0.5152 34 1.6165 0.3898 1.6165 1.2714
No log 0.5455 36 1.6119 0.3802 1.6119 1.2696
No log 0.5758 38 1.6222 0.3871 1.6222 1.2736
No log 0.6061 40 1.6679 0.3876 1.6679 1.2915
No log 0.6364 42 1.6242 0.4154 1.6242 1.2744
No log 0.6667 44 1.3872 0.5 1.3872 1.1778
No log 0.6970 46 1.2852 0.5152 1.2852 1.1337
No log 0.7273 48 1.3391 0.5109 1.3391 1.1572
No log 0.7576 50 1.5419 0.4571 1.5419 1.2417
No log 0.7879 52 1.6945 0.3586 1.6945 1.3017
No log 0.8182 54 1.5372 0.4247 1.5372 1.2398
No log 0.8485 56 1.4236 0.4667 1.4236 1.1931
No log 0.8788 58 1.2876 0.4507 1.2876 1.1347
No log 0.9091 60 1.2596 0.5 1.2596 1.1223
No log 0.9394 62 1.2385 0.4925 1.2385 1.1129
No log 0.9697 64 1.0886 0.5735 1.0886 1.0434
No log 1.0 66 1.0400 0.6528 1.0400 1.0198
No log 1.0303 68 1.0154 0.6301 1.0154 1.0077
No log 1.0606 70 0.9240 0.6029 0.9240 0.9613
No log 1.0909 72 1.3143 0.5075 1.3143 1.1464
No log 1.1212 74 1.5997 0.3947 1.5997 1.2648
No log 1.1515 76 1.3405 0.5298 1.3405 1.1578
No log 1.1818 78 0.9323 0.5985 0.9323 0.9656
No log 1.2121 80 0.8579 0.6269 0.8579 0.9262
No log 1.2424 82 0.9663 0.6164 0.9663 0.9830
No log 1.2727 84 1.3040 0.5556 1.3040 1.1419
No log 1.3030 86 1.7299 0.4944 1.7299 1.3153
No log 1.3333 88 1.5899 0.5414 1.5899 1.2609
No log 1.3636 90 1.1596 0.6125 1.1596 1.0768
No log 1.3939 92 1.0697 0.6081 1.0697 1.0343
No log 1.4242 94 1.1415 0.5571 1.1415 1.0684
No log 1.4545 96 1.3648 0.4604 1.3648 1.1682
No log 1.4848 98 1.4343 0.368 1.4343 1.1976
No log 1.5152 100 1.0787 0.5469 1.0787 1.0386
No log 1.5455 102 0.8961 0.6260 0.8961 0.9466
No log 1.5758 104 0.8562 0.6619 0.8562 0.9253
No log 1.6061 106 0.8040 0.7310 0.8040 0.8966
No log 1.6364 108 0.7737 0.7297 0.7737 0.8796
No log 1.6667 110 0.8075 0.6711 0.8075 0.8986
No log 1.6970 112 0.8551 0.6879 0.8551 0.9247
No log 1.7273 114 0.9283 0.6829 0.9283 0.9635
No log 1.7576 116 1.0040 0.6951 1.0040 1.0020
No log 1.7879 118 1.0210 0.6133 1.0210 1.0105
No log 1.8182 120 0.9802 0.6345 0.9802 0.9900
No log 1.8485 122 0.9530 0.6277 0.9530 0.9762
No log 1.8788 124 0.9514 0.6119 0.9514 0.9754
No log 1.9091 126 0.9964 0.6212 0.9964 0.9982
No log 1.9394 128 1.0143 0.5426 1.0143 1.0071
No log 1.9697 130 1.0011 0.5827 1.0011 1.0006
No log 2.0 132 1.1298 0.5079 1.1298 1.0629
No log 2.0303 134 1.6224 0.2901 1.6224 1.2737
No log 2.0606 136 1.9659 0.2059 1.9659 1.4021
No log 2.0909 138 1.3839 0.4234 1.3839 1.1764
No log 2.1212 140 0.9767 0.5970 0.9767 0.9883
No log 2.1515 142 0.8298 0.6950 0.8298 0.9109
No log 2.1818 144 0.8383 0.7114 0.8383 0.9156
No log 2.2121 146 0.7960 0.7613 0.7960 0.8922
No log 2.2424 148 0.7531 0.6892 0.7531 0.8678
No log 2.2727 150 0.7576 0.6712 0.7576 0.8704
No log 2.3030 152 0.8322 0.6928 0.8322 0.9122
No log 2.3333 154 1.1663 0.6364 1.1663 1.0799
No log 2.3636 156 1.0966 0.6347 1.0966 1.0472
No log 2.3939 158 0.7819 0.7044 0.7819 0.8843
No log 2.4242 160 0.7250 0.7547 0.7250 0.8515
No log 2.4545 162 0.7402 0.7296 0.7402 0.8603
No log 2.4848 164 0.7278 0.7484 0.7278 0.8531
No log 2.5152 166 0.7253 0.7097 0.7253 0.8516
No log 2.5455 168 0.8220 0.7152 0.8220 0.9066
No log 2.5758 170 1.0008 0.6782 1.0008 1.0004
No log 2.6061 172 1.0320 0.6782 1.0320 1.0159
No log 2.6364 174 0.9085 0.6788 0.9085 0.9532
No log 2.6667 176 0.8532 0.6667 0.8532 0.9237
No log 2.6970 178 0.9193 0.6620 0.9193 0.9588
No log 2.7273 180 0.9603 0.6324 0.9603 0.9799
No log 2.7576 182 0.9352 0.6377 0.9352 0.9670
No log 2.7879 184 0.8254 0.6667 0.8254 0.9085
No log 2.8182 186 0.7894 0.7027 0.7894 0.8885
No log 2.8485 188 0.8812 0.6918 0.8812 0.9387
No log 2.8788 190 0.9611 0.6788 0.9611 0.9804
No log 2.9091 192 0.9394 0.675 0.9394 0.9692
No log 2.9394 194 0.8806 0.6986 0.8806 0.9384
No log 2.9697 196 0.8735 0.6573 0.8735 0.9346
No log 3.0 198 0.8924 0.6575 0.8924 0.9447
No log 3.0303 200 1.0123 0.6536 1.0123 1.0061
No log 3.0606 202 1.0105 0.6541 1.0105 1.0052
No log 3.0909 204 0.8682 0.6797 0.8682 0.9318
No log 3.1212 206 0.8207 0.7162 0.8207 0.9059
No log 3.1515 208 0.8923 0.6577 0.8923 0.9446
No log 3.1818 210 0.9973 0.6531 0.9973 0.9986
No log 3.2121 212 1.3377 0.5101 1.3377 1.1566
No log 3.2424 214 1.6420 0.4487 1.6420 1.2814
No log 3.2727 216 1.5398 0.4177 1.5398 1.2409
No log 3.3030 218 1.1568 0.5960 1.1568 1.0755
No log 3.3333 220 0.9223 0.6933 0.9223 0.9604
No log 3.3636 222 0.8497 0.6887 0.8497 0.9218
No log 3.3939 224 0.7466 0.7190 0.7466 0.8640
No log 3.4242 226 0.6925 0.7568 0.6925 0.8322
No log 3.4545 228 0.7031 0.7568 0.7031 0.8385
No log 3.4848 230 0.7348 0.7682 0.7348 0.8572
No log 3.5152 232 0.7815 0.6986 0.7815 0.8840
No log 3.5455 234 0.8121 0.6577 0.8121 0.9012
No log 3.5758 236 0.7902 0.6800 0.7902 0.8890
No log 3.6061 238 0.8444 0.6624 0.8444 0.9189
No log 3.6364 240 1.0897 0.6667 1.0897 1.0439
No log 3.6667 242 1.0939 0.6631 1.0939 1.0459
No log 3.6970 244 0.8361 0.7073 0.8361 0.9144
No log 3.7273 246 0.7248 0.7632 0.7248 0.8514
No log 3.7576 248 0.8194 0.7027 0.8194 0.9052
No log 3.7879 250 0.8602 0.6714 0.8602 0.9275
No log 3.8182 252 0.8712 0.6853 0.8712 0.9334
No log 3.8485 254 0.8978 0.6849 0.8978 0.9475
No log 3.8788 256 0.9700 0.6486 0.9700 0.9849
No log 3.9091 258 0.9541 0.6434 0.9541 0.9768
No log 3.9394 260 0.9079 0.6429 0.9079 0.9529
No log 3.9697 262 0.7770 0.7368 0.7770 0.8815
No log 4.0 264 0.7197 0.7484 0.7197 0.8484
No log 4.0303 266 0.7624 0.7561 0.7624 0.8731
No log 4.0606 268 0.7627 0.7561 0.7627 0.8733
No log 4.0909 270 0.7798 0.7702 0.7798 0.8831
No log 4.1212 272 0.7394 0.7484 0.7394 0.8599
No log 4.1515 274 0.7491 0.7368 0.7491 0.8655
No log 4.1818 276 0.7773 0.7712 0.7773 0.8816
No log 4.2121 278 0.8857 0.7114 0.8857 0.9411
No log 4.2424 280 1.0639 0.6375 1.0639 1.0314
No log 4.2727 282 1.0116 0.6375 1.0116 1.0058
No log 4.3030 284 0.8289 0.7097 0.8289 0.9105
No log 4.3333 286 0.7522 0.7383 0.7522 0.8673
No log 4.3636 288 0.7914 0.6577 0.7914 0.8896
No log 4.3939 290 0.8020 0.6849 0.8020 0.8955
No log 4.4242 292 0.8091 0.6809 0.8091 0.8995
No log 4.4545 294 0.8918 0.6667 0.8918 0.9443
No log 4.4848 296 0.9666 0.6711 0.9666 0.9832
No log 4.5152 298 0.9713 0.6849 0.9713 0.9855
No log 4.5455 300 0.9017 0.6849 0.9017 0.9496
No log 4.5758 302 0.7716 0.6892 0.7716 0.8784
No log 4.6061 304 0.7398 0.7417 0.7398 0.8601
No log 4.6364 306 0.8001 0.7143 0.8001 0.8945
No log 4.6667 308 0.8618 0.6887 0.8618 0.9284
No log 4.6970 310 0.8852 0.6887 0.8852 0.9409
No log 4.7273 312 0.7678 0.72 0.7678 0.8762
No log 4.7576 314 0.6861 0.7383 0.6861 0.8283
No log 4.7879 316 0.7292 0.7172 0.7292 0.8539
No log 4.8182 318 0.7685 0.7183 0.7685 0.8766
No log 4.8485 320 0.7858 0.6901 0.7858 0.8864
No log 4.8788 322 0.7361 0.7467 0.7361 0.8580
No log 4.9091 324 0.6822 0.7632 0.6822 0.8260
No log 4.9394 326 0.6576 0.7190 0.6576 0.8109
No log 4.9697 328 0.6598 0.7190 0.6598 0.8123
No log 5.0 330 0.6694 0.7662 0.6694 0.8182
No log 5.0303 332 0.7299 0.7949 0.7299 0.8543
No log 5.0606 334 0.8826 0.6928 0.8826 0.9395
No log 5.0909 336 0.9617 0.6974 0.9617 0.9807
No log 5.1212 338 0.9369 0.6974 0.9369 0.9679
No log 5.1515 340 0.8621 0.7105 0.8621 0.9285
No log 5.1818 342 0.7366 0.7483 0.7366 0.8583
No log 5.2121 344 0.6967 0.7183 0.6967 0.8347
No log 5.2424 346 0.7972 0.6615 0.7972 0.8928
No log 5.2727 348 0.8812 0.6080 0.8812 0.9387
No log 5.3030 350 0.9979 0.6471 0.9979 0.9989
No log 5.3333 352 1.1201 0.5547 1.1201 1.0583
No log 5.3636 354 1.0749 0.5986 1.0749 1.0368
No log 5.3939 356 0.9008 0.7075 0.9008 0.9491
No log 5.4242 358 0.7572 0.6950 0.7572 0.8702
No log 5.4545 360 0.7749 0.6667 0.7749 0.8803
No log 5.4848 362 0.8069 0.6857 0.8069 0.8983
No log 5.5152 364 0.8738 0.6761 0.8738 0.9348
No log 5.5455 366 0.8997 0.6667 0.8997 0.9485
No log 5.5758 368 0.8625 0.6759 0.8625 0.9287
No log 5.6061 370 0.7648 0.7172 0.7648 0.8745
No log 5.6364 372 0.7542 0.7133 0.7542 0.8684
No log 5.6667 374 0.7621 0.6715 0.7621 0.8730
No log 5.6970 376 0.7669 0.7133 0.7669 0.8757
No log 5.7273 378 0.8157 0.75 0.8157 0.9032
No log 5.7576 380 0.8908 0.7179 0.8908 0.9438
No log 5.7879 382 0.8487 0.7273 0.8487 0.9212
No log 5.8182 384 0.7643 0.7632 0.7643 0.8742
No log 5.8485 386 0.7427 0.7376 0.7427 0.8618
No log 5.8788 388 0.7322 0.7376 0.7322 0.8557
No log 5.9091 390 0.7287 0.7050 0.7287 0.8537
No log 5.9394 392 0.7063 0.7143 0.7063 0.8404
No log 5.9697 394 0.7044 0.7376 0.7044 0.8393
No log 6.0 396 0.7047 0.7338 0.7047 0.8395
No log 6.0303 398 0.7148 0.7059 0.7148 0.8454
No log 6.0606 400 0.7206 0.7246 0.7206 0.8489
No log 6.0909 402 0.7138 0.7413 0.7138 0.8449
No log 6.1212 404 0.7023 0.7534 0.7023 0.8380
No log 6.1515 406 0.6773 0.7651 0.6773 0.8230
No log 6.1818 408 0.6725 0.7703 0.6725 0.8201
No log 6.2121 410 0.7014 0.7273 0.7014 0.8375
No log 6.2424 412 0.7242 0.7222 0.7242 0.8510
No log 6.2727 414 0.7897 0.6993 0.7897 0.8887
No log 6.3030 416 0.8927 0.6887 0.8927 0.9448
No log 6.3333 418 0.9086 0.6622 0.9086 0.9532
No log 6.3636 420 0.8123 0.6849 0.8123 0.9013
No log 6.3939 422 0.7222 0.7101 0.7222 0.8498
No log 6.4242 424 0.6901 0.7 0.6901 0.8308
No log 6.4545 426 0.6607 0.7222 0.6607 0.8129
No log 6.4848 428 0.6418 0.7347 0.6418 0.8011
No log 6.5152 430 0.6235 0.7297 0.6235 0.7896
No log 6.5455 432 0.6123 0.7448 0.6123 0.7825
No log 6.5758 434 0.6041 0.7671 0.6041 0.7772
No log 6.6061 436 0.6303 0.7974 0.6303 0.7939
No log 6.6364 438 0.6679 0.7632 0.6679 0.8173
No log 6.6667 440 0.6612 0.7632 0.6612 0.8131
No log 6.6970 442 0.6984 0.7550 0.6984 0.8357
No log 6.7273 444 0.6978 0.7413 0.6978 0.8354
No log 6.7576 446 0.7804 0.7092 0.7804 0.8834
No log 6.7879 448 0.8665 0.6906 0.8665 0.9309
No log 6.8182 450 0.9724 0.6619 0.9724 0.9861
No log 6.8485 452 1.0490 0.5899 1.0490 1.0242
No log 6.8788 454 0.9805 0.6620 0.9805 0.9902
No log 6.9091 456 0.8280 0.6761 0.8280 0.9099
No log 6.9394 458 0.7276 0.7413 0.7276 0.8530
No log 6.9697 460 0.7297 0.7639 0.7297 0.8542
No log 7.0 462 0.7287 0.7586 0.7287 0.8536
No log 7.0303 464 0.7446 0.7808 0.7446 0.8629
No log 7.0606 466 0.7372 0.7671 0.7372 0.8586
No log 7.0909 468 0.7272 0.7671 0.7272 0.8528
No log 7.1212 470 0.7506 0.7413 0.7506 0.8664
No log 7.1515 472 0.8445 0.6939 0.8445 0.9189
No log 7.1818 474 0.8790 0.6712 0.8790 0.9375
No log 7.2121 476 0.8360 0.7133 0.8360 0.9143
No log 7.2424 478 0.7569 0.7413 0.7569 0.8700
No log 7.2727 480 0.7409 0.7671 0.7409 0.8608
No log 7.3030 482 0.7347 0.75 0.7347 0.8571
No log 7.3333 484 0.7896 0.7172 0.7896 0.8886
No log 7.3636 486 0.8034 0.7172 0.8034 0.8964
No log 7.3939 488 0.7603 0.7260 0.7603 0.8720
No log 7.4242 490 0.7120 0.7755 0.7120 0.8438
No log 7.4545 492 0.6992 0.7733 0.6992 0.8362
No log 7.4848 494 0.6975 0.7815 0.6975 0.8352
No log 7.5152 496 0.7282 0.7582 0.7282 0.8533
No log 7.5455 498 0.7540 0.7368 0.7540 0.8683
0.4454 7.5758 500 0.7349 0.7582 0.7349 0.8573
0.4454 7.6061 502 0.6931 0.7703 0.6931 0.8325
0.4454 7.6364 504 0.6888 0.7619 0.6888 0.8300
0.4454 7.6667 506 0.6842 0.7619 0.6842 0.8272
0.4454 7.6970 508 0.6736 0.7808 0.6736 0.8207
0.4454 7.7273 510 0.6780 0.7891 0.6780 0.8234
0.4454 7.7576 512 0.7038 0.7808 0.7038 0.8389
0.4454 7.7879 514 0.7295 0.7413 0.7295 0.8541
0.4454 7.8182 516 0.7688 0.7234 0.7688 0.8768
0.4454 7.8485 518 0.8091 0.7007 0.8091 0.8995
0.4454 7.8788 520 0.8212 0.6567 0.8212 0.9062
0.4454 7.9091 522 0.8269 0.7183 0.8269 0.9094

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k14_task1_organization

Finetuned
(4019)
this model