ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k12_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7193
  • Qwk: 0.7194
  • Mse: 0.7193
  • Rmse: 0.8481

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0357 2 6.9054 0.0058 6.9054 2.6278
No log 0.0714 4 4.5406 0.0506 4.5406 2.1309
No log 0.1071 6 3.5936 -0.0659 3.5936 1.8957
No log 0.1429 8 3.0735 -0.0513 3.0735 1.7531
No log 0.1786 10 2.0908 0.0522 2.0908 1.4460
No log 0.2143 12 1.8017 0.1165 1.8017 1.3423
No log 0.25 14 1.8527 0.1538 1.8527 1.3611
No log 0.2857 16 2.0048 0.2162 2.0048 1.4159
No log 0.3214 18 2.1829 0.1452 2.1829 1.4775
No log 0.3571 20 2.2085 0.1364 2.2085 1.4861
No log 0.3929 22 1.9433 0.2222 1.9433 1.3940
No log 0.4286 24 1.8355 0.2707 1.8355 1.3548
No log 0.4643 26 2.3055 0.3333 2.3055 1.5184
No log 0.5 28 3.3880 0.1956 3.3880 1.8406
No log 0.5357 30 3.9165 0.1935 3.9165 1.9790
No log 0.5714 32 2.5241 0.3300 2.5241 1.5887
No log 0.6071 34 1.5030 0.4459 1.5030 1.2260
No log 0.6429 36 1.4242 0.3968 1.4242 1.1934
No log 0.6786 38 1.3511 0.3710 1.3511 1.1624
No log 0.7143 40 1.4475 0.4690 1.4475 1.2031
No log 0.75 42 1.8005 0.4699 1.8005 1.3418
No log 0.7857 44 2.0523 0.4094 2.0523 1.4326
No log 0.8214 46 2.4961 0.2703 2.4961 1.5799
No log 0.8571 48 2.7812 0.2447 2.7812 1.6677
No log 0.8929 50 2.3264 0.2469 2.3264 1.5253
No log 0.9286 52 1.6048 0.4148 1.6048 1.2668
No log 0.9643 54 1.3995 0.3636 1.3995 1.1830
No log 1.0 56 1.4383 0.3770 1.4383 1.1993
No log 1.0357 58 1.4317 0.4 1.4317 1.1965
No log 1.0714 60 1.4454 0.4341 1.4454 1.2023
No log 1.1071 62 1.3756 0.4567 1.3756 1.1729
No log 1.1429 64 1.2211 0.4567 1.2211 1.1050
No log 1.1786 66 1.2638 0.5315 1.2638 1.1242
No log 1.2143 68 1.9588 0.4541 1.9588 1.3996
No log 1.25 70 2.6483 0.3519 2.6483 1.6274
No log 1.2857 72 2.4219 0.4019 2.4219 1.5562
No log 1.3214 74 1.6194 0.5435 1.6194 1.2726
No log 1.3571 76 1.5622 0.5532 1.5622 1.2499
No log 1.3929 78 1.4017 0.6087 1.4017 1.1839
No log 1.4286 80 1.1709 0.6286 1.1709 1.0821
No log 1.4643 82 1.0388 0.6053 1.0388 1.0192
No log 1.5 84 1.0355 0.5811 1.0355 1.0176
No log 1.5357 86 1.1161 0.6258 1.1161 1.0565
No log 1.5714 88 1.0711 0.6184 1.0711 1.0350
No log 1.6071 90 1.1679 0.5890 1.1679 1.0807
No log 1.6429 92 1.1268 0.6065 1.1268 1.0615
No log 1.6786 94 1.0517 0.5946 1.0517 1.0255
No log 1.7143 96 1.1889 0.5584 1.1889 1.0904
No log 1.75 98 1.2707 0.5629 1.2707 1.1273
No log 1.7857 100 0.9713 0.6405 0.9713 0.9856
No log 1.8214 102 0.8819 0.7042 0.8819 0.9391
No log 1.8571 104 0.9239 0.5882 0.9239 0.9612
No log 1.8929 106 0.9514 0.6143 0.9514 0.9754
No log 1.9286 108 0.9991 0.5972 0.9991 0.9995
No log 1.9643 110 1.0670 0.5987 1.0670 1.0330
No log 2.0 112 1.0780 0.6076 1.0780 1.0383
No log 2.0357 114 0.9057 0.6753 0.9057 0.9517
No log 2.0714 116 0.8328 0.6968 0.8328 0.9126
No log 2.1071 118 0.8552 0.6918 0.8552 0.9248
No log 2.1429 120 0.9999 0.6707 0.9999 0.9999
No log 2.1786 122 0.9849 0.6707 0.9849 0.9924
No log 2.2143 124 0.8094 0.7355 0.8094 0.8997
No log 2.25 126 0.8332 0.6849 0.8332 0.9128
No log 2.2857 128 0.9211 0.6434 0.9211 0.9597
No log 2.3214 130 0.8303 0.6849 0.8303 0.9112
No log 2.3571 132 0.7535 0.7602 0.7535 0.8680
No log 2.3929 134 1.0710 0.7010 1.0710 1.0349
No log 2.4286 136 1.0581 0.7077 1.0581 1.0286
No log 2.4643 138 0.7917 0.7619 0.7917 0.8898
No log 2.5 140 0.7172 0.7711 0.7172 0.8468
No log 2.5357 142 0.7356 0.7484 0.7356 0.8577
No log 2.5714 144 0.7471 0.7561 0.7471 0.8644
No log 2.6071 146 0.7878 0.7586 0.7878 0.8876
No log 2.6429 148 0.7775 0.7471 0.7775 0.8817
No log 2.6786 150 0.7261 0.7613 0.7261 0.8521
No log 2.7143 152 0.7677 0.7632 0.7677 0.8762
No log 2.75 154 0.7661 0.7632 0.7661 0.8753
No log 2.7857 156 0.7408 0.7613 0.7408 0.8607
No log 2.8214 158 0.7532 0.7516 0.7532 0.8678
No log 2.8571 160 0.7461 0.7516 0.7461 0.8638
No log 2.8929 162 0.7085 0.7516 0.7085 0.8417
No log 2.9286 164 0.7183 0.7403 0.7183 0.8475
No log 2.9643 166 0.7180 0.7403 0.7180 0.8474
No log 3.0 168 0.7307 0.7417 0.7307 0.8548
No log 3.0357 170 0.7179 0.7417 0.7179 0.8473
No log 3.0714 172 0.7203 0.7613 0.7203 0.8487
No log 3.1071 174 0.7386 0.7467 0.7386 0.8594
No log 3.1429 176 0.7507 0.7237 0.7507 0.8664
No log 3.1786 178 0.7783 0.6950 0.7783 0.8822
No log 3.2143 180 0.7640 0.7397 0.7640 0.8741
No log 3.25 182 0.7772 0.7417 0.7772 0.8816
No log 3.2857 184 0.7639 0.7568 0.7639 0.8740
No log 3.3214 186 0.7482 0.7324 0.7482 0.8650
No log 3.3571 188 0.7467 0.7234 0.7467 0.8641
No log 3.3929 190 0.7786 0.7042 0.7786 0.8824
No log 3.4286 192 0.7507 0.75 0.7507 0.8665
No log 3.4643 194 0.7766 0.7397 0.7766 0.8813
No log 3.5 196 0.8309 0.7347 0.8309 0.9115
No log 3.5357 198 0.9139 0.6761 0.9139 0.9560
No log 3.5714 200 1.0082 0.6131 1.0082 1.0041
No log 3.6071 202 0.8807 0.6812 0.8807 0.9385
No log 3.6429 204 0.8395 0.6950 0.8395 0.9162
No log 3.6786 206 0.8797 0.5778 0.8797 0.9379
No log 3.7143 208 0.9568 0.5672 0.9568 0.9782
No log 3.75 210 0.8891 0.6176 0.8891 0.9429
No log 3.7857 212 0.8421 0.7133 0.8421 0.9177
No log 3.8214 214 0.8121 0.7448 0.8121 0.9012
No log 3.8571 216 1.0086 0.625 1.0086 1.0043
No log 3.8929 218 1.2872 0.6196 1.2872 1.1346
No log 3.9286 220 0.9823 0.6228 0.9823 0.9911
No log 3.9643 222 0.7718 0.7792 0.7718 0.8785
No log 4.0 224 0.7684 0.7568 0.7684 0.8766
No log 4.0357 226 0.7851 0.7517 0.7851 0.8861
No log 4.0714 228 0.7941 0.7417 0.7941 0.8911
No log 4.1071 230 0.8167 0.7432 0.8167 0.9037
No log 4.1429 232 0.9113 0.7075 0.9113 0.9546
No log 4.1786 234 0.9419 0.6986 0.9419 0.9705
No log 4.2143 236 0.8583 0.7310 0.8583 0.9265
No log 4.25 238 0.8426 0.7639 0.8426 0.9179
No log 4.2857 240 0.8270 0.7639 0.8270 0.9094
No log 4.3214 242 0.7879 0.7639 0.7879 0.8876
No log 4.3571 244 0.7499 0.7724 0.7499 0.8659
No log 4.3929 246 0.7500 0.7671 0.7500 0.8660
No log 4.4286 248 0.8037 0.6944 0.8037 0.8965
No log 4.4643 250 0.7781 0.6993 0.7781 0.8821
No log 4.5 252 0.6863 0.7724 0.6863 0.8284
No log 4.5357 254 0.7277 0.7436 0.7277 0.8530
No log 4.5714 256 0.7401 0.7436 0.7401 0.8603
No log 4.6071 258 0.6720 0.7742 0.6720 0.8198
No log 4.6429 260 0.7037 0.7310 0.7037 0.8388
No log 4.6786 262 0.8000 0.6713 0.8000 0.8944
No log 4.7143 264 0.8550 0.6667 0.8550 0.9246
No log 4.75 266 0.8308 0.6853 0.8308 0.9115
No log 4.7857 268 0.7885 0.6897 0.7885 0.8880
No log 4.8214 270 0.8264 0.7034 0.8264 0.9091
No log 4.8571 272 0.8215 0.7034 0.8215 0.9063
No log 4.8929 274 0.7350 0.7075 0.7350 0.8573
No log 4.9286 276 0.7062 0.7483 0.7062 0.8404
No log 4.9643 278 0.7160 0.7310 0.7160 0.8462
No log 5.0 280 0.7586 0.7273 0.7586 0.8710
No log 5.0357 282 0.9165 0.6571 0.9165 0.9574
No log 5.0714 284 1.0710 0.5735 1.0710 1.0349
No log 5.1071 286 1.0157 0.6131 1.0157 1.0078
No log 5.1429 288 0.8336 0.7260 0.8336 0.9130
No log 5.1786 290 0.7541 0.7347 0.7541 0.8684
No log 5.2143 292 0.7501 0.7550 0.7501 0.8661
No log 5.25 294 0.7570 0.7383 0.7570 0.8700
No log 5.2857 296 0.7834 0.7682 0.7834 0.8851
No log 5.3214 298 0.7651 0.7467 0.7651 0.8747
No log 5.3571 300 0.7574 0.7582 0.7574 0.8703
No log 5.3929 302 0.8035 0.7248 0.8035 0.8964
No log 5.4286 304 0.9212 0.7143 0.9212 0.9598
No log 5.4643 306 0.9628 0.6861 0.9628 0.9812
No log 5.5 308 0.8764 0.7133 0.8764 0.9361
No log 5.5357 310 0.7974 0.7133 0.7974 0.8930
No log 5.5714 312 0.7864 0.7297 0.7864 0.8868
No log 5.6071 314 0.8754 0.6410 0.8754 0.9356
No log 5.6429 316 0.9994 0.6460 0.9994 0.9997
No log 5.6786 318 0.8906 0.6582 0.8906 0.9437
No log 5.7143 320 0.7888 0.7059 0.7888 0.8881
No log 5.75 322 0.8028 0.7376 0.8028 0.8960
No log 5.7857 324 0.8748 0.7286 0.8748 0.9353
No log 5.8214 326 0.8923 0.7194 0.8923 0.9446
No log 5.8571 328 0.7818 0.7815 0.7818 0.8842
No log 5.8929 330 0.7793 0.7296 0.7793 0.8828
No log 5.9286 332 0.8622 0.7283 0.8622 0.9286
No log 5.9643 334 0.8682 0.7374 0.8682 0.9318
No log 6.0 336 0.7366 0.7329 0.7366 0.8582
No log 6.0357 338 0.8638 0.6713 0.8638 0.9294
No log 6.0714 340 0.9596 0.6571 0.9596 0.9796
No log 6.1071 342 0.9391 0.6618 0.9391 0.9691
No log 6.1429 344 0.8603 0.6950 0.8603 0.9275
No log 6.1786 346 0.7925 0.7483 0.7925 0.8902
No log 6.2143 348 0.7316 0.7703 0.7316 0.8554
No log 6.25 350 0.7341 0.7619 0.7341 0.8568
No log 6.2857 352 0.7220 0.7534 0.7220 0.8497
No log 6.3214 354 0.7220 0.76 0.7220 0.8497
No log 6.3571 356 0.7945 0.7320 0.7945 0.8914
No log 6.3929 358 0.7891 0.7152 0.7891 0.8883
No log 6.4286 360 0.7294 0.7682 0.7294 0.8541
No log 6.4643 362 0.7156 0.7619 0.7156 0.8460
No log 6.5 364 0.6968 0.7712 0.6968 0.8347
No log 6.5357 366 0.7065 0.7712 0.7065 0.8406
No log 6.5714 368 0.7564 0.7389 0.7564 0.8697
No log 6.6071 370 0.7899 0.7226 0.7899 0.8888
No log 6.6429 372 0.7643 0.7389 0.7643 0.8742
No log 6.6786 374 0.7319 0.7547 0.7319 0.8555
No log 6.7143 376 0.7621 0.7317 0.7621 0.8730
No log 6.75 378 0.7906 0.7284 0.7906 0.8891
No log 6.7857 380 0.8250 0.7296 0.8250 0.9083
No log 6.8214 382 0.8426 0.7089 0.8426 0.9179
No log 6.8571 384 0.9022 0.7089 0.9022 0.9499
No log 6.8929 386 0.9679 0.7152 0.9679 0.9838
No log 6.9286 388 0.9642 0.7108 0.9642 0.9820
No log 6.9643 390 0.8988 0.7439 0.8988 0.9481
No log 7.0 392 0.7450 0.7468 0.7450 0.8631
No log 7.0357 394 0.6956 0.7702 0.6956 0.8340
No log 7.0714 396 0.6663 0.7901 0.6663 0.8163
No log 7.1071 398 0.6389 0.8 0.6389 0.7993
No log 7.1429 400 0.6473 0.7882 0.6473 0.8045
No log 7.1786 402 0.6380 0.7953 0.6380 0.7988
No log 7.2143 404 0.6249 0.7692 0.6249 0.7905
No log 7.25 406 0.7324 0.7273 0.7324 0.8558
No log 7.2857 408 0.8964 0.6950 0.8964 0.9468
No log 7.3214 410 0.9245 0.7143 0.9245 0.9615
No log 7.3571 412 0.9072 0.6812 0.9072 0.9525
No log 7.3929 414 0.8305 0.6950 0.8305 0.9113
No log 7.4286 416 0.8049 0.6812 0.8049 0.8972
No log 7.4643 418 0.7804 0.6950 0.7804 0.8834
No log 7.5 420 0.7485 0.6950 0.7485 0.8652
No log 7.5357 422 0.7313 0.7310 0.7313 0.8552
No log 7.5714 424 0.7021 0.7532 0.7021 0.8379
No log 7.6071 426 0.7008 0.7613 0.7008 0.8371
No log 7.6429 428 0.7063 0.7643 0.7063 0.8404
No log 7.6786 430 0.7334 0.7821 0.7334 0.8564
No log 7.7143 432 0.7555 0.7347 0.7555 0.8692
No log 7.75 434 0.7488 0.7248 0.7488 0.8653
No log 7.7857 436 0.7283 0.7432 0.7283 0.8534
No log 7.8214 438 0.7482 0.7712 0.7482 0.8650
No log 7.8571 440 0.7292 0.7712 0.7292 0.8539
No log 7.8929 442 0.6798 0.7534 0.6798 0.8245
No log 7.9286 444 0.7049 0.75 0.7049 0.8396
No log 7.9643 446 0.8112 0.6809 0.8112 0.9006
No log 8.0 448 0.8886 0.6857 0.8886 0.9426
No log 8.0357 450 0.7999 0.7042 0.7999 0.8944
No log 8.0714 452 0.7170 0.7413 0.7170 0.8468
No log 8.1071 454 0.6951 0.7568 0.6951 0.8337
No log 8.1429 456 0.7052 0.7639 0.7052 0.8397
No log 8.1786 458 0.6996 0.7742 0.6996 0.8364
No log 8.2143 460 0.7219 0.7564 0.7219 0.8496
No log 8.25 462 0.7590 0.7 0.7590 0.8712
No log 8.2857 464 0.7883 0.6835 0.7883 0.8879
No log 8.3214 466 0.8280 0.6752 0.8280 0.9099
No log 8.3571 468 0.8621 0.6581 0.8621 0.9285
No log 8.3929 470 0.9093 0.6460 0.9093 0.9536
No log 8.4286 472 0.8769 0.6788 0.8769 0.9364
No log 8.4643 474 0.8696 0.7168 0.8696 0.9325
No log 8.5 476 0.8837 0.7273 0.8837 0.9400
No log 8.5357 478 0.8855 0.7273 0.8855 0.9410
No log 8.5714 480 0.6720 0.7425 0.6720 0.8198
No log 8.6071 482 0.5956 0.8050 0.5956 0.7718
No log 8.6429 484 0.5882 0.7785 0.5882 0.7669
No log 8.6786 486 0.6152 0.7671 0.6152 0.7844
No log 8.7143 488 0.6475 0.7619 0.6475 0.8047
No log 8.75 490 0.7236 0.7273 0.7236 0.8507
No log 8.7857 492 0.7986 0.7042 0.7986 0.8936
No log 8.8214 494 0.8156 0.7286 0.8156 0.9031
No log 8.8571 496 0.7983 0.7286 0.7983 0.8935
No log 8.8929 498 0.8021 0.7286 0.8021 0.8956
0.4126 8.9286 500 0.7892 0.7273 0.7892 0.8884
0.4126 8.9643 502 0.7878 0.7172 0.7878 0.8876
0.4126 9.0 504 0.7463 0.7347 0.7463 0.8639
0.4126 9.0357 506 0.7322 0.7397 0.7322 0.8557
0.4126 9.0714 508 0.7596 0.7183 0.7596 0.8715
0.4126 9.1071 510 0.7434 0.7376 0.7434 0.8622
0.4126 9.1429 512 0.7342 0.7465 0.7342 0.8569
0.4126 9.1786 514 0.7422 0.7465 0.7422 0.8615
0.4126 9.2143 516 0.7162 0.7465 0.7162 0.8463
0.4126 9.25 518 0.7065 0.6861 0.7065 0.8405
0.4126 9.2857 520 0.7416 0.6765 0.7416 0.8611
0.4126 9.3214 522 0.7913 0.6567 0.7913 0.8896
0.4126 9.3571 524 0.7580 0.6815 0.7580 0.8706
0.4126 9.3929 526 0.7193 0.7194 0.7193 0.8481

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k12_task1_organization

Finetuned
(4019)
this model