ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k13_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0155
  • Qwk: 0.5
  • Mse: 1.0155
  • Rmse: 1.0077

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0333 2 6.6660 0.0308 6.6660 2.5819
No log 0.0667 4 4.2832 0.1004 4.2832 2.0696
No log 0.1 6 2.9218 0.0755 2.9218 1.7093
No log 0.1333 8 2.6895 0.0745 2.6895 1.6400
No log 0.1667 10 2.2188 0.0709 2.2188 1.4896
No log 0.2 12 1.8785 0.1739 1.8785 1.3706
No log 0.2333 14 1.7554 0.1651 1.7554 1.3249
No log 0.2667 16 1.7661 0.1754 1.7661 1.3289
No log 0.3 18 1.9226 0.1803 1.9226 1.3866
No log 0.3333 20 1.8854 0.1356 1.8854 1.3731
No log 0.3667 22 1.9151 0.1724 1.9151 1.3839
No log 0.4 24 2.2116 0.0730 2.2116 1.4872
No log 0.4333 26 2.1737 0.1926 2.1737 1.4743
No log 0.4667 28 1.7968 0.3622 1.7968 1.3404
No log 0.5 30 1.9526 0.3158 1.9526 1.3974
No log 0.5333 32 2.6169 0.0645 2.6169 1.6177
No log 0.5667 34 3.6060 0.0769 3.6060 1.8989
No log 0.6 36 3.3094 0.0795 3.3094 1.8192
No log 0.6333 38 2.5103 0.1074 2.5103 1.5844
No log 0.6667 40 1.8023 0.3415 1.8023 1.3425
No log 0.7 42 1.5202 0.3361 1.5202 1.2330
No log 0.7333 44 1.6216 0.3361 1.6216 1.2734
No log 0.7667 46 1.8147 0.3175 1.8147 1.3471
No log 0.8 48 1.9606 0.3231 1.9606 1.4002
No log 0.8333 50 1.9894 0.3231 1.9894 1.4105
No log 0.8667 52 2.0013 0.2576 2.0013 1.4147
No log 0.9 54 2.0138 0.2074 2.0138 1.4191
No log 0.9333 56 1.8402 0.3125 1.8402 1.3566
No log 0.9667 58 1.6360 0.2602 1.6360 1.2791
No log 1.0 60 1.5528 0.3025 1.5528 1.2461
No log 1.0333 62 1.5675 0.3167 1.5675 1.2520
No log 1.0667 64 1.6110 0.3387 1.6110 1.2693
No log 1.1 66 1.6416 0.3040 1.6416 1.2812
No log 1.1333 68 1.5544 0.3810 1.5544 1.2468
No log 1.1667 70 1.4576 0.3667 1.4576 1.2073
No log 1.2 72 1.4441 0.3969 1.4441 1.2017
No log 1.2333 74 1.6011 0.3942 1.6011 1.2654
No log 1.2667 76 2.1036 0.2857 2.1036 1.4504
No log 1.3 78 2.1601 0.2857 2.1601 1.4697
No log 1.3333 80 1.6636 0.3830 1.6636 1.2898
No log 1.3667 82 1.2617 0.4762 1.2617 1.1232
No log 1.4 84 1.3040 0.4516 1.3040 1.1419
No log 1.4333 86 1.4038 0.4032 1.4038 1.1848
No log 1.4667 88 1.4402 0.4094 1.4402 1.2001
No log 1.5 90 1.4435 0.4496 1.4435 1.2015
No log 1.5333 92 1.6899 0.3741 1.6899 1.3000
No log 1.5667 94 1.9833 0.3265 1.9833 1.4083
No log 1.6 96 2.0616 0.2895 2.0616 1.4358
No log 1.6333 98 1.8263 0.3357 1.8263 1.3514
No log 1.6667 100 1.5069 0.4265 1.5069 1.2276
No log 1.7 102 1.1181 0.5954 1.1181 1.0574
No log 1.7333 104 1.0150 0.5970 1.0150 1.0075
No log 1.7667 106 0.9340 0.6715 0.9340 0.9664
No log 1.8 108 0.9571 0.6074 0.9571 0.9783
No log 1.8333 110 1.2306 0.5132 1.2306 1.1093
No log 1.8667 112 1.1857 0.5217 1.1857 1.0889
No log 1.9 114 0.9588 0.5816 0.9588 0.9792
No log 1.9333 116 1.0662 0.6119 1.0662 1.0326
No log 1.9667 118 1.2879 0.5116 1.2879 1.1349
No log 2.0 120 1.1617 0.5037 1.1617 1.0778
No log 2.0333 122 1.1364 0.4925 1.1364 1.0660
No log 2.0667 124 1.1192 0.5735 1.1192 1.0579
No log 2.1 126 1.0068 0.6528 1.0068 1.0034
No log 2.1333 128 0.9303 0.6620 0.9303 0.9645
No log 2.1667 130 0.8225 0.7172 0.8225 0.9069
No log 2.2 132 0.8396 0.6853 0.8396 0.9163
No log 2.2333 134 0.8076 0.6901 0.8076 0.8986
No log 2.2667 136 0.7920 0.6897 0.7920 0.8900
No log 2.3 138 0.8012 0.6809 0.8012 0.8951
No log 2.3333 140 0.8109 0.6571 0.8109 0.9005
No log 2.3667 142 0.8169 0.6522 0.8169 0.9038
No log 2.4 144 0.7925 0.6763 0.7925 0.8902
No log 2.4333 146 0.7940 0.6857 0.7940 0.8911
No log 2.4667 148 0.7703 0.7172 0.7703 0.8777
No log 2.5 150 0.7884 0.7297 0.7884 0.8879
No log 2.5333 152 0.8835 0.6324 0.8835 0.9400
No log 2.5667 154 0.9740 0.6222 0.9740 0.9869
No log 2.6 156 0.9720 0.5865 0.9720 0.9859
No log 2.6333 158 1.0248 0.5926 1.0248 1.0123
No log 2.6667 160 1.0683 0.5758 1.0683 1.0336
No log 2.7 162 1.0698 0.5082 1.0698 1.0343
No log 2.7333 164 1.1672 0.5528 1.1672 1.0804
No log 2.7667 166 1.1723 0.5323 1.1723 1.0827
No log 2.8 168 1.0152 0.5625 1.0152 1.0076
No log 2.8333 170 0.9819 0.5846 0.9819 0.9909
No log 2.8667 172 0.9923 0.5821 0.9923 0.9961
No log 2.9 174 0.9060 0.6119 0.9060 0.9519
No log 2.9333 176 0.9127 0.6370 0.9127 0.9554
No log 2.9667 178 0.9206 0.6015 0.9206 0.9595
No log 3.0 180 1.1725 0.5915 1.1725 1.0828
No log 3.0333 182 1.3343 0.5467 1.3343 1.1551
No log 3.0667 184 1.1186 0.5816 1.1186 1.0577
No log 3.1 186 0.9009 0.6269 0.9009 0.9491
No log 3.1333 188 1.0467 0.6061 1.0467 1.0231
No log 3.1667 190 1.0682 0.6260 1.0682 1.0335
No log 3.2 192 0.9749 0.6154 0.9749 0.9874
No log 3.2333 194 1.0343 0.5606 1.0343 1.0170
No log 3.2667 196 1.1061 0.5373 1.1061 1.0517
No log 3.3 198 1.0405 0.5496 1.0405 1.0200
No log 3.3333 200 1.0175 0.5312 1.0175 1.0087
No log 3.3667 202 1.0895 0.6260 1.0895 1.0438
No log 3.4 204 1.1232 0.6269 1.1232 1.0598
No log 3.4333 206 1.0148 0.6519 1.0148 1.0074
No log 3.4667 208 0.9536 0.5909 0.9536 0.9765
No log 3.5 210 1.0057 0.5079 1.0057 1.0029
No log 3.5333 212 0.9963 0.48 0.9963 0.9981
No log 3.5667 214 0.9832 0.6154 0.9832 0.9916
No log 3.6 216 1.0379 0.6308 1.0379 1.0188
No log 3.6333 218 1.0060 0.6142 1.0060 1.0030
No log 3.6667 220 0.9858 0.6094 0.9858 0.9929
No log 3.7 222 0.9370 0.5802 0.9370 0.9680
No log 3.7333 224 0.9252 0.5692 0.9252 0.9619
No log 3.7667 226 0.9415 0.5909 0.9415 0.9703
No log 3.8 228 0.9547 0.5802 0.9547 0.9771
No log 3.8333 230 0.9605 0.5802 0.9605 0.9800
No log 3.8667 232 1.0129 0.5865 1.0129 1.0064
No log 3.9 234 1.1381 0.5564 1.1381 1.0668
No log 3.9333 236 1.1251 0.5564 1.1251 1.0607
No log 3.9667 238 1.0195 0.5238 1.0195 1.0097
No log 4.0 240 0.9879 0.5323 0.9879 0.9939
No log 4.0333 242 0.9801 0.5397 0.9801 0.9900
No log 4.0667 244 1.0138 0.5238 1.0138 1.0069
No log 4.1 246 1.0037 0.5238 1.0037 1.0019
No log 4.1333 248 1.0031 0.5354 1.0031 1.0016
No log 4.1667 250 0.9945 0.5649 0.9945 0.9972
No log 4.2 252 1.0379 0.5469 1.0379 1.0188
No log 4.2333 254 1.1285 0.5 1.1285 1.0623
No log 4.2667 256 1.1430 0.4918 1.1430 1.0691
No log 4.3 258 1.1287 0.5238 1.1287 1.0624
No log 4.3333 260 1.1348 0.5238 1.1348 1.0653
No log 4.3667 262 1.1763 0.5323 1.1763 1.0846
No log 4.4 264 1.3561 0.4667 1.3561 1.1645
No log 4.4333 266 1.2985 0.4553 1.2985 1.1395
No log 4.4667 268 1.1246 0.5556 1.1246 1.0605
No log 4.5 270 0.9099 0.6462 0.9099 0.9539
No log 4.5333 272 0.8916 0.6462 0.8916 0.9442
No log 4.5667 274 0.9077 0.6357 0.9077 0.9527
No log 4.6 276 0.9586 0.6154 0.9586 0.9791
No log 4.6333 278 1.0214 0.5692 1.0214 1.0107
No log 4.6667 280 0.9489 0.6260 0.9489 0.9741
No log 4.7 282 0.7721 0.7234 0.7721 0.8787
No log 4.7333 284 0.7044 0.7891 0.7044 0.8393
No log 4.7667 286 0.7209 0.7808 0.7209 0.8491
No log 4.8 288 0.7704 0.7050 0.7704 0.8777
No log 4.8333 290 0.8342 0.6565 0.8342 0.9134
No log 4.8667 292 0.8759 0.6615 0.8759 0.9359
No log 4.9 294 0.8709 0.6512 0.8709 0.9332
No log 4.9333 296 0.8836 0.6565 0.8836 0.9400
No log 4.9667 298 0.9009 0.6165 0.9009 0.9492
No log 5.0 300 0.9440 0.6165 0.9440 0.9716
No log 5.0333 302 1.0167 0.5781 1.0167 1.0083
No log 5.0667 304 1.0017 0.5781 1.0017 1.0009
No log 5.1 306 0.9080 0.6 0.9080 0.9529
No log 5.1333 308 0.8625 0.5625 0.8625 0.9287
No log 5.1667 310 0.9260 0.5496 0.9260 0.9623
No log 5.2 312 0.9051 0.5426 0.9051 0.9514
No log 5.2333 314 0.8531 0.6269 0.8531 0.9236
No log 5.2667 316 0.8427 0.6815 0.8427 0.9180
No log 5.3 318 0.8693 0.6519 0.8693 0.9323
No log 5.3333 320 0.8222 0.7153 0.8222 0.9068
No log 5.3667 322 0.7959 0.6765 0.7959 0.8921
No log 5.4 324 0.8040 0.6765 0.8040 0.8967
No log 5.4333 326 0.8081 0.6715 0.8081 0.8989
No log 5.4667 328 0.8661 0.6567 0.8661 0.9306
No log 5.5 330 0.9073 0.6565 0.9073 0.9525
No log 5.5333 332 0.8770 0.6466 0.8770 0.9365
No log 5.5667 334 0.8788 0.5354 0.8788 0.9374
No log 5.6 336 0.8802 0.5581 0.8802 0.9382
No log 5.6333 338 0.9056 0.6 0.9056 0.9516
No log 5.6667 340 0.9228 0.6047 0.9228 0.9606
No log 5.7 342 0.9175 0.6 0.9175 0.9578
No log 5.7333 344 0.9198 0.6107 0.9198 0.9591
No log 5.7667 346 0.9456 0.6047 0.9456 0.9724
No log 5.8 348 0.9814 0.5271 0.9814 0.9907
No log 5.8333 350 1.0090 0.5079 1.0090 1.0045
No log 5.8667 352 0.9931 0.496 0.9931 0.9966
No log 5.9 354 0.9602 0.5645 0.9602 0.9799
No log 5.9333 356 0.9637 0.5738 0.9637 0.9817
No log 5.9667 358 1.0017 0.5528 1.0017 1.0008
No log 6.0 360 0.9614 0.5806 0.9614 0.9805
No log 6.0333 362 0.9275 0.5806 0.9275 0.9631
No log 6.0667 364 0.8832 0.5238 0.8832 0.9398
No log 6.1 366 0.8991 0.5714 0.8991 0.9482
No log 6.1333 368 0.9098 0.5312 0.9098 0.9538
No log 6.1667 370 0.9324 0.5197 0.9324 0.9656
No log 6.2 372 0.9558 0.512 0.9558 0.9776
No log 6.2333 374 0.9626 0.512 0.9626 0.9811
No log 6.2667 376 0.9695 0.5354 0.9695 0.9847
No log 6.3 378 0.9890 0.5440 0.9890 0.9945
No log 6.3333 380 0.9977 0.5484 0.9977 0.9988
No log 6.3667 382 1.0030 0.544 1.0030 1.0015
No log 6.4 384 0.9870 0.544 0.9870 0.9935
No log 6.4333 386 0.9630 0.5556 0.9630 0.9813
No log 6.4667 388 0.9551 0.5366 0.9551 0.9773
No log 6.5 390 0.9259 0.5366 0.9259 0.9623
No log 6.5333 392 0.8928 0.5827 0.8928 0.9449
No log 6.5667 394 0.8568 0.5891 0.8568 0.9257
No log 6.6 396 0.7975 0.6418 0.7975 0.8931
No log 6.6333 398 0.7951 0.6617 0.7951 0.8917
No log 6.6667 400 0.8015 0.6466 0.8015 0.8953
No log 6.7 402 0.7942 0.6812 0.7942 0.8912
No log 6.7333 404 0.7905 0.6950 0.7905 0.8891
No log 6.7667 406 0.7804 0.7183 0.7804 0.8834
No log 6.8 408 0.8118 0.6466 0.8118 0.9010
No log 6.8333 410 0.8566 0.5873 0.8566 0.9255
No log 6.8667 412 0.8763 0.5984 0.8763 0.9361
No log 6.9 414 0.8886 0.5556 0.8886 0.9426
No log 6.9333 416 0.9121 0.5954 0.9121 0.9550
No log 6.9667 418 0.8678 0.6061 0.8678 0.9315
No log 7.0 420 0.7998 0.6715 0.7998 0.8943
No log 7.0333 422 0.7657 0.6619 0.7657 0.8751
No log 7.0667 424 0.7450 0.7133 0.7450 0.8631
No log 7.1 426 0.7547 0.7042 0.7547 0.8688
No log 7.1333 428 0.7952 0.6906 0.7952 0.8918
No log 7.1667 430 0.8572 0.6667 0.8572 0.9259
No log 7.2 432 0.9192 0.5 0.9192 0.9588
No log 7.2333 434 0.9423 0.5397 0.9423 0.9707
No log 7.2667 436 0.9794 0.544 0.9794 0.9896
No log 7.3 438 0.9668 0.4878 0.9668 0.9833
No log 7.3333 440 0.9802 0.5246 0.9802 0.9900
No log 7.3667 442 1.0083 0.5469 1.0083 1.0041
No log 7.4 444 0.9431 0.5669 0.9431 0.9711
No log 7.4333 446 0.8855 0.6061 0.8855 0.9410
No log 7.4667 448 0.8593 0.6047 0.8593 0.9270
No log 7.5 450 0.8716 0.6212 0.8716 0.9336
No log 7.5333 452 0.9168 0.6087 0.9168 0.9575
No log 7.5667 454 0.9447 0.5942 0.9447 0.9720
No log 7.6 456 0.8773 0.6377 0.8773 0.9366
No log 7.6333 458 0.8259 0.6617 0.8259 0.9088
No log 7.6667 460 0.8627 0.6412 0.8627 0.9288
No log 7.7 462 0.9119 0.5645 0.9119 0.9549
No log 7.7333 464 0.9583 0.5124 0.9583 0.9789
No log 7.7667 466 0.9969 0.4706 0.9969 0.9985
No log 7.8 468 1.0137 0.5041 1.0137 1.0068
No log 7.8333 470 0.9976 0.5692 0.9976 0.9988
No log 7.8667 472 0.9290 0.6260 0.9290 0.9638
No log 7.9 474 0.8577 0.6260 0.8577 0.9261
No log 7.9333 476 0.8245 0.6866 0.8245 0.9080
No log 7.9667 478 0.8235 0.7059 0.8235 0.9075
No log 8.0 480 0.8048 0.7153 0.8048 0.8971
No log 8.0333 482 0.7999 0.7246 0.7999 0.8944
No log 8.0667 484 0.9345 0.5672 0.9345 0.9667
No log 8.1 486 1.0193 0.5344 1.0193 1.0096
No log 8.1333 488 0.9797 0.5606 0.9797 0.9898
No log 8.1667 490 0.8914 0.5954 0.8914 0.9442
No log 8.2 492 0.8559 0.6716 0.8559 0.9251
No log 8.2333 494 0.8677 0.6515 0.8677 0.9315
No log 8.2667 496 0.8992 0.5891 0.8992 0.9483
No log 8.3 498 0.9484 0.5512 0.9484 0.9739
0.4076 8.3333 500 0.9811 0.5512 0.9811 0.9905
0.4076 8.3667 502 1.0052 0.5 1.0052 1.0026
0.4076 8.4 504 0.9897 0.5 0.9897 0.9949
0.4076 8.4333 506 0.9566 0.5714 0.9566 0.9781
0.4076 8.4667 508 0.9186 0.6202 0.9186 0.9584
0.4076 8.5 510 0.8825 0.6212 0.8825 0.9394
0.4076 8.5333 512 0.8808 0.6357 0.8808 0.9385
0.4076 8.5667 514 0.8867 0.625 0.8867 0.9416
0.4076 8.6 516 0.8920 0.5556 0.8920 0.9445
0.4076 8.6333 518 0.9512 0.5197 0.9512 0.9753
0.4076 8.6667 520 0.9722 0.5625 0.9722 0.9860
0.4076 8.7 522 0.9483 0.5397 0.9483 0.9738
0.4076 8.7333 524 0.9520 0.5556 0.9520 0.9757
0.4076 8.7667 526 0.9376 0.56 0.9376 0.9683
0.4076 8.8 528 0.9640 0.5581 0.9640 0.9818
0.4076 8.8333 530 0.9537 0.5366 0.9537 0.9766
0.4076 8.8667 532 0.9632 0.5161 0.9632 0.9814
0.4076 8.9 534 1.0117 0.4839 1.0117 1.0058
0.4076 8.9333 536 1.0592 0.4426 1.0592 1.0292
0.4076 8.9667 538 1.0155 0.5 1.0155 1.0077

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k13_task1_organization

Finetuned
(4019)
this model