ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k12_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.1753
  • Qwk: 0.6282
  • Mse: 1.1753
  • Rmse: 1.0841

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0222 2 7.1885 0.0174 7.1885 2.6811
No log 0.0444 4 4.9873 0.0509 4.9873 2.2332
No log 0.0667 6 3.1660 0.0952 3.1660 1.7793
No log 0.0889 8 2.1915 0.1399 2.1915 1.4804
No log 0.1111 10 1.9308 0.2131 1.9308 1.3895
No log 0.1333 12 1.7346 0.2655 1.7346 1.3170
No log 0.1556 14 1.7085 0.1538 1.7085 1.3071
No log 0.1778 16 2.2706 -0.0168 2.2706 1.5069
No log 0.2 18 2.4314 -0.1280 2.4314 1.5593
No log 0.2222 20 1.6977 0.1538 1.6977 1.3030
No log 0.2444 22 1.1009 0.512 1.1009 1.0492
No log 0.2667 24 1.1384 0.4628 1.1384 1.0670
No log 0.2889 26 1.2714 0.5736 1.2714 1.1276
No log 0.3111 28 1.7383 0.2581 1.7383 1.3185
No log 0.3333 30 1.8101 0.1524 1.8101 1.3454
No log 0.3556 32 1.9821 0.1111 1.9821 1.4079
No log 0.3778 34 1.6957 0.1887 1.6957 1.3022
No log 0.4 36 1.4148 0.1905 1.4148 1.1894
No log 0.4222 38 1.2408 0.3966 1.2408 1.1139
No log 0.4444 40 1.2368 0.5113 1.2368 1.1121
No log 0.4667 42 1.1948 0.4885 1.1948 1.0931
No log 0.4889 44 1.2775 0.3276 1.2775 1.1302
No log 0.5111 46 1.3625 0.3214 1.3625 1.1673
No log 0.5333 48 1.4436 0.3276 1.4436 1.2015
No log 0.5556 50 1.2276 0.3793 1.2276 1.1080
No log 0.5778 52 0.9721 0.5522 0.9721 0.9859
No log 0.6 54 0.9525 0.6286 0.9525 0.9760
No log 0.6222 56 1.0052 0.6370 1.0052 1.0026
No log 0.6444 58 1.0963 0.6324 1.0963 1.0470
No log 0.6667 60 1.1140 0.6074 1.1140 1.0555
No log 0.6889 62 1.0925 0.5954 1.0925 1.0452
No log 0.7111 64 1.1594 0.5954 1.1594 1.0768
No log 0.7333 66 1.1295 0.5954 1.1295 1.0628
No log 0.7556 68 1.0497 0.5954 1.0497 1.0246
No log 0.7778 70 0.9703 0.5954 0.9703 0.9850
No log 0.8 72 1.0386 0.6212 1.0386 1.0191
No log 0.8222 74 0.9705 0.6466 0.9705 0.9851
No log 0.8444 76 0.7706 0.6618 0.7706 0.8778
No log 0.8667 78 0.7163 0.7034 0.7163 0.8463
No log 0.8889 80 0.8042 0.7484 0.8042 0.8968
No log 0.9111 82 0.7860 0.7673 0.7860 0.8865
No log 0.9333 84 0.7276 0.6980 0.7276 0.8530
No log 0.9556 86 0.7955 0.6571 0.7955 0.8919
No log 0.9778 88 0.9942 0.6 0.9942 0.9971
No log 1.0 90 1.0651 0.5850 1.0651 1.0320
No log 1.0222 92 1.1591 0.5772 1.1591 1.0766
No log 1.0444 94 1.1523 0.5455 1.1523 1.0735
No log 1.0667 96 1.1112 0.5594 1.1112 1.0541
No log 1.0889 98 1.1416 0.5352 1.1416 1.0685
No log 1.1111 100 1.1933 0.5652 1.1933 1.0924
No log 1.1333 102 1.1986 0.5612 1.1986 1.0948
No log 1.1556 104 1.1136 0.5594 1.1136 1.0553
No log 1.1778 106 1.0964 0.6309 1.0964 1.0471
No log 1.2 108 1.1372 0.6289 1.1372 1.0664
No log 1.2222 110 0.9754 0.6667 0.9754 0.9876
No log 1.2444 112 0.8027 0.6621 0.8027 0.8959
No log 1.2667 114 0.7873 0.6806 0.7873 0.8873
No log 1.2889 116 0.8239 0.6928 0.8239 0.9077
No log 1.3111 118 0.7433 0.7027 0.7433 0.8621
No log 1.3333 120 0.7386 0.6761 0.7386 0.8594
No log 1.3556 122 0.8064 0.6980 0.8064 0.8980
No log 1.3778 124 0.7767 0.6897 0.7767 0.8813
No log 1.4 126 0.7097 0.6993 0.7097 0.8424
No log 1.4222 128 0.7954 0.6928 0.7954 0.8919
No log 1.4444 130 1.2040 0.5799 1.2040 1.0973
No log 1.4667 132 1.3111 0.5442 1.3111 1.1450
No log 1.4889 134 1.1172 0.5634 1.1172 1.0570
No log 1.5111 136 0.9715 0.6241 0.9715 0.9856
No log 1.5333 138 0.9017 0.6933 0.9017 0.9496
No log 1.5556 140 0.9446 0.7229 0.9446 0.9719
No log 1.5778 142 1.0194 0.6951 1.0194 1.0096
No log 1.6 144 1.2076 0.6272 1.2076 1.0989
No log 1.6222 146 1.0771 0.6316 1.0771 1.0379
No log 1.6444 148 0.9838 0.6405 0.9838 0.9919
No log 1.6667 150 1.1205 0.6497 1.1205 1.0585
No log 1.6889 152 1.0308 0.6234 1.0308 1.0153
No log 1.7111 154 0.7688 0.7067 0.7688 0.8768
No log 1.7333 156 0.7207 0.7484 0.7207 0.8489
No log 1.7556 158 0.7379 0.7273 0.7379 0.8590
No log 1.7778 160 0.8968 0.6710 0.8968 0.9470
No log 1.8 162 1.0967 0.6480 1.0967 1.0473
No log 1.8222 164 1.1965 0.6392 1.1965 1.0939
No log 1.8444 166 1.0672 0.6667 1.0672 1.0331
No log 1.8667 168 0.7571 0.7791 0.7571 0.8701
No log 1.8889 170 0.5857 0.8353 0.5857 0.7653
No log 1.9111 172 0.6378 0.7927 0.6378 0.7986
No log 1.9333 174 0.6578 0.7778 0.6578 0.8111
No log 1.9556 176 0.8378 0.6897 0.8378 0.9153
No log 1.9778 178 0.9888 0.6423 0.9888 0.9944
No log 2.0 180 1.0929 0.5571 1.0929 1.0454
No log 2.0222 182 1.0644 0.5931 1.0644 1.0317
No log 2.0444 184 1.1643 0.5679 1.1643 1.0790
No log 2.0667 186 1.2190 0.5965 1.2190 1.1041
No log 2.0889 188 1.2755 0.5839 1.2755 1.1294
No log 2.1111 190 1.2211 0.5369 1.2211 1.1050
No log 2.1333 192 1.1549 0.5342 1.1549 1.0747
No log 2.1556 194 1.1595 0.5823 1.1595 1.0768
No log 2.1778 196 1.0873 0.6296 1.0873 1.0427
No log 2.2 198 0.9726 0.6887 0.9726 0.9862
No log 2.2222 200 0.8778 0.6331 0.8778 0.9369
No log 2.2444 202 0.7580 0.6761 0.7580 0.8706
No log 2.2667 204 0.7725 0.6897 0.7725 0.8789
No log 2.2889 206 0.9424 0.6579 0.9424 0.9708
No log 2.3111 208 1.0334 0.6588 1.0334 1.0166
No log 2.3333 210 0.8414 0.6746 0.8414 0.9173
No log 2.3556 212 0.6709 0.7952 0.6709 0.8191
No log 2.3778 214 0.7090 0.7702 0.7090 0.8420
No log 2.4 216 0.8646 0.6860 0.8646 0.9298
No log 2.4222 218 1.0876 0.6667 1.0876 1.0429
No log 2.4444 220 1.0818 0.6545 1.0818 1.0401
No log 2.4667 222 0.9655 0.6623 0.9655 0.9826
No log 2.4889 224 0.7332 0.7034 0.7332 0.8563
No log 2.5111 226 0.6746 0.7222 0.6746 0.8213
No log 2.5333 228 0.6443 0.7742 0.6443 0.8027
No log 2.5556 230 0.6468 0.8121 0.6468 0.8043
No log 2.5778 232 0.8613 0.7487 0.8613 0.9281
No log 2.6 234 1.0704 0.6857 1.0704 1.0346
No log 2.6222 236 1.0257 0.6536 1.0257 1.0128
No log 2.6444 238 0.8883 0.6757 0.8883 0.9425
No log 2.6667 240 0.8514 0.6475 0.8514 0.9227
No log 2.6889 242 0.9047 0.6522 0.9047 0.9511
No log 2.7111 244 0.9199 0.6331 0.9199 0.9591
No log 2.7333 246 0.8678 0.6906 0.8678 0.9315
No log 2.7556 248 0.8793 0.6667 0.8793 0.9377
No log 2.7778 250 0.8703 0.6944 0.8703 0.9329
No log 2.8 252 0.8775 0.6944 0.8775 0.9368
No log 2.8222 254 0.9758 0.6914 0.9758 0.9878
No log 2.8444 256 1.0241 0.675 1.0241 1.0120
No log 2.8667 258 0.9504 0.6944 0.9504 0.9749
No log 2.8889 260 0.9378 0.6809 0.9378 0.9684
No log 2.9111 262 0.9921 0.6389 0.9921 0.9960
No log 2.9333 264 1.0467 0.6490 1.0467 1.0231
No log 2.9556 266 0.9050 0.6974 0.9050 0.9513
No log 2.9778 268 0.6740 0.7448 0.6740 0.8210
No log 3.0 270 0.6064 0.7922 0.6064 0.7787
No log 3.0222 272 0.6303 0.7792 0.6303 0.7939
No log 3.0444 274 0.8179 0.7261 0.8179 0.9044
No log 3.0667 276 1.0426 0.6375 1.0426 1.0211
No log 3.0889 278 1.0513 0.6375 1.0513 1.0254
No log 3.1111 280 0.8816 0.6667 0.8816 0.9389
No log 3.1333 282 0.7480 0.6515 0.7480 0.8649
No log 3.1556 284 0.7568 0.6767 0.7568 0.8700
No log 3.1778 286 0.8220 0.6618 0.8220 0.9067
No log 3.2 288 0.8629 0.6892 0.8629 0.9289
No log 3.2222 290 0.8294 0.6887 0.8294 0.9107
No log 3.2444 292 0.9536 0.6709 0.9536 0.9765
No log 3.2667 294 1.0766 0.6552 1.0766 1.0376
No log 3.2889 296 0.9583 0.6708 0.9583 0.9789
No log 3.3111 298 0.8380 0.6389 0.8380 0.9154
No log 3.3333 300 0.8153 0.6475 0.8153 0.9029
No log 3.3556 302 0.8149 0.6525 0.8149 0.9027
No log 3.3778 304 0.7848 0.7027 0.7848 0.8859
No log 3.4 306 0.8200 0.7108 0.8200 0.9056
No log 3.4222 308 0.9080 0.675 0.9080 0.9529
No log 3.4444 310 0.9690 0.675 0.9690 0.9844
No log 3.4667 312 0.8292 0.7097 0.8292 0.9106
No log 3.4889 314 0.7359 0.7625 0.7359 0.8579
No log 3.5111 316 0.7065 0.7417 0.7065 0.8405
No log 3.5333 318 0.6467 0.7320 0.6467 0.8041
No log 3.5556 320 0.6607 0.7403 0.6607 0.8129
No log 3.5778 322 0.8185 0.7528 0.8185 0.9047
No log 3.6 324 0.9659 0.6821 0.9659 0.9828
No log 3.6222 326 0.8542 0.6711 0.8542 0.9243
No log 3.6444 328 0.8286 0.6812 0.8286 0.9103
No log 3.6667 330 0.8796 0.6423 0.8796 0.9379
No log 3.6889 332 0.9739 0.5926 0.9739 0.9869
No log 3.7111 334 0.9490 0.6901 0.9490 0.9742
No log 3.7333 336 0.9276 0.6623 0.9276 0.9631
No log 3.7556 338 0.9002 0.7117 0.9002 0.9488
No log 3.7778 340 0.7949 0.7160 0.7949 0.8916
No log 3.8 342 0.6567 0.7702 0.6567 0.8104
No log 3.8222 344 0.6295 0.7643 0.6295 0.7934
No log 3.8444 346 0.6939 0.7683 0.6939 0.8330
No log 3.8667 348 0.8822 0.7006 0.8822 0.9393
No log 3.8889 350 0.9622 0.6927 0.9622 0.9809
No log 3.9111 352 0.8466 0.7126 0.8466 0.9201
No log 3.9333 354 0.7740 0.7037 0.7740 0.8798
No log 3.9556 356 0.6694 0.7582 0.6694 0.8182
No log 3.9778 358 0.6576 0.7815 0.6576 0.8109
No log 4.0 360 0.7197 0.7368 0.7197 0.8484
No log 4.0222 362 0.7432 0.7226 0.7432 0.8621
No log 4.0444 364 0.8510 0.7152 0.8510 0.9225
No log 4.0667 366 0.8768 0.7152 0.8768 0.9364
No log 4.0889 368 0.8382 0.6918 0.8382 0.9156
No log 4.1111 370 0.8647 0.6800 0.8647 0.9299
No log 4.1333 372 0.8785 0.6803 0.8785 0.9373
No log 4.1556 374 0.9189 0.6712 0.9189 0.9586
No log 4.1778 376 0.9578 0.6667 0.9578 0.9787
No log 4.2 378 0.9443 0.6667 0.9443 0.9717
No log 4.2222 380 0.8436 0.6803 0.8436 0.9185
No log 4.2444 382 0.7609 0.7059 0.7609 0.8723
No log 4.2667 384 0.7326 0.7143 0.7326 0.8559
No log 4.2889 386 0.7612 0.7059 0.7612 0.8725
No log 4.3111 388 0.8175 0.7059 0.8175 0.9041
No log 4.3333 390 0.8378 0.7059 0.8378 0.9153
No log 4.3556 392 0.8200 0.6667 0.8200 0.9055
No log 4.3778 394 0.7505 0.7383 0.7505 0.8663
No log 4.4 396 0.7333 0.7383 0.7333 0.8563
No log 4.4222 398 0.7837 0.7020 0.7837 0.8852
No log 4.4444 400 0.8961 0.6792 0.8961 0.9466
No log 4.4667 402 0.9326 0.6914 0.9326 0.9657
No log 4.4889 404 0.8225 0.6525 0.8225 0.9069
No log 4.5111 406 0.7446 0.6906 0.7446 0.8629
No log 4.5333 408 0.7540 0.6812 0.7540 0.8683
No log 4.5556 410 0.8235 0.6812 0.8235 0.9075
No log 4.5778 412 0.8757 0.6522 0.8757 0.9358
No log 4.6 414 0.9843 0.6711 0.9843 0.9921
No log 4.6222 416 1.0469 0.6790 1.0469 1.0232
No log 4.6444 418 1.0051 0.6790 1.0051 1.0025
No log 4.6667 420 0.8554 0.6846 0.8554 0.9249
No log 4.6889 422 0.8035 0.7162 0.8035 0.8964
No log 4.7111 424 0.8560 0.6939 0.8560 0.9252
No log 4.7333 426 0.8616 0.7516 0.8616 0.9282
No log 4.7556 428 0.8452 0.7421 0.8452 0.9193
No log 4.7778 430 0.8030 0.7439 0.8030 0.8961
No log 4.8 432 0.6389 0.7792 0.6389 0.7993
No log 4.8222 434 0.5837 0.7632 0.5837 0.7640
No log 4.8444 436 0.5893 0.7742 0.5893 0.7677
No log 4.8667 438 0.6270 0.7904 0.6270 0.7918
No log 4.8889 440 0.7931 0.7650 0.7931 0.8905
No log 4.9111 442 0.9028 0.7232 0.9028 0.9502
No log 4.9333 444 0.9042 0.7045 0.9042 0.9509
No log 4.9556 446 0.8513 0.7086 0.8513 0.9227
No log 4.9778 448 0.6816 0.7347 0.6816 0.8256
No log 5.0 450 0.6279 0.75 0.6279 0.7924
No log 5.0222 452 0.6486 0.7568 0.6486 0.8054
No log 5.0444 454 0.7621 0.7215 0.7621 0.8730
No log 5.0667 456 0.8376 0.7093 0.8376 0.9152
No log 5.0889 458 0.7521 0.7412 0.7521 0.8672
No log 5.1111 460 0.7093 0.7383 0.7093 0.8422
No log 5.1333 462 0.7466 0.7383 0.7466 0.8640
No log 5.1556 464 0.8131 0.6708 0.8131 0.9017
No log 5.1778 466 0.7626 0.7075 0.7626 0.8733
No log 5.2 468 0.7741 0.7152 0.7741 0.8798
No log 5.2222 470 0.8861 0.7086 0.8861 0.9413
No log 5.2444 472 0.9742 0.6556 0.9742 0.9870
No log 5.2667 474 1.0539 0.6480 1.0539 1.0266
No log 5.2889 476 0.9439 0.6590 0.9439 0.9715
No log 5.3111 478 0.7632 0.7075 0.7632 0.8736
No log 5.3333 480 0.6798 0.6667 0.6798 0.8245
No log 5.3556 482 0.6756 0.6565 0.6756 0.8219
No log 5.3778 484 0.7193 0.7162 0.7193 0.8481
No log 5.4 486 0.8515 0.6864 0.8515 0.9228
No log 5.4222 488 0.9618 0.6932 0.9618 0.9807
No log 5.4444 490 1.0990 0.6592 1.0990 1.0483
No log 5.4667 492 1.1127 0.6629 1.1127 1.0549
No log 5.4889 494 0.9544 0.6792 0.9544 0.9769
No log 5.5111 496 0.8352 0.6667 0.8352 0.9139
No log 5.5333 498 0.8239 0.6308 0.8239 0.9077
0.45 5.5556 500 0.8908 0.6569 0.8908 0.9438
0.45 5.5778 502 1.0814 0.6538 1.0814 1.0399
0.45 5.6 504 1.2131 0.6310 1.2131 1.1014
0.45 5.6222 506 1.1788 0.6629 1.1788 1.0857
0.45 5.6444 508 1.0703 0.6667 1.0703 1.0346
0.45 5.6667 510 0.8696 0.6573 0.8696 0.9325
0.45 5.6889 512 0.7872 0.6569 0.7872 0.8872
0.45 5.7111 514 0.8097 0.6569 0.8097 0.8998
0.45 5.7333 516 0.8909 0.6569 0.8909 0.9439
0.45 5.7556 518 1.0639 0.6174 1.0639 1.0314
0.45 5.7778 520 1.1753 0.6282 1.1753 1.0841

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k12_task1_organization

Finetuned
(4023)
this model