ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k12_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0057
  • Qwk: 0.2984
  • Mse: 1.0057
  • Rmse: 1.0029

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0333 2 4.8596 -0.0240 4.8596 2.2044
No log 0.0667 4 2.7633 0.0306 2.7633 1.6623
No log 0.1 6 1.8276 0.0198 1.8276 1.3519
No log 0.1333 8 1.8372 -0.0468 1.8372 1.3554
No log 0.1667 10 1.9311 -0.0073 1.9311 1.3896
No log 0.2 12 1.5494 0.0334 1.5494 1.2447
No log 0.2333 14 1.3585 0.0140 1.3585 1.1655
No log 0.2667 16 1.2600 0.1327 1.2600 1.1225
No log 0.3 18 1.5945 0.0936 1.5945 1.2627
No log 0.3333 20 1.6546 0.1055 1.6546 1.2863
No log 0.3667 22 1.6621 0.1795 1.6621 1.2892
No log 0.4 24 1.3998 0.1972 1.3998 1.1831
No log 0.4333 26 1.2241 0.1951 1.2241 1.1064
No log 0.4667 28 1.1805 0.2044 1.1805 1.0865
No log 0.5 30 1.1680 0.2191 1.1680 1.0808
No log 0.5333 32 1.3536 0.2431 1.3536 1.1634
No log 0.5667 34 1.3552 0.2062 1.3552 1.1641
No log 0.6 36 1.2574 0.2333 1.2574 1.1213
No log 0.6333 38 1.4326 0.2155 1.4326 1.1969
No log 0.6667 40 1.8761 0.2495 1.8761 1.3697
No log 0.7 42 1.8554 0.2826 1.8554 1.3621
No log 0.7333 44 1.3697 0.2474 1.3697 1.1703
No log 0.7667 46 1.2397 0.2881 1.2397 1.1134
No log 0.8 48 1.1917 0.2520 1.1917 1.0917
No log 0.8333 50 1.0485 0.3373 1.0485 1.0240
No log 0.8667 52 1.0363 0.4388 1.0363 1.0180
No log 0.9 54 1.0637 0.2366 1.0637 1.0314
No log 0.9333 56 1.2334 0.2201 1.2334 1.1106
No log 0.9667 58 1.3253 0.2201 1.3253 1.1512
No log 1.0 60 1.3532 0.2627 1.3532 1.1633
No log 1.0333 62 1.2955 0.2677 1.2955 1.1382
No log 1.0667 64 1.4101 0.2653 1.4101 1.1875
No log 1.1 66 1.5886 0.2808 1.5886 1.2604
No log 1.1333 68 1.8183 0.2742 1.8183 1.3484
No log 1.1667 70 1.8390 0.2632 1.8390 1.3561
No log 1.2 72 1.7432 0.2795 1.7432 1.3203
No log 1.2333 74 1.5174 0.3857 1.5174 1.2318
No log 1.2667 76 1.3209 0.4554 1.3209 1.1493
No log 1.3 78 1.1473 0.3545 1.1473 1.0711
No log 1.3333 80 1.0089 0.4676 1.0089 1.0044
No log 1.3667 82 0.9800 0.4459 0.9800 0.9899
No log 1.4 84 0.9225 0.4705 0.9225 0.9605
No log 1.4333 86 0.9722 0.3568 0.9722 0.9860
No log 1.4667 88 0.9042 0.4123 0.9042 0.9509
No log 1.5 90 0.9247 0.4305 0.9247 0.9616
No log 1.5333 92 1.2739 0.4423 1.2739 1.1287
No log 1.5667 94 1.3673 0.4313 1.3673 1.1693
No log 1.6 96 1.3050 0.4568 1.3050 1.1424
No log 1.6333 98 1.0476 0.4518 1.0476 1.0235
No log 1.6667 100 0.9493 0.5487 0.9493 0.9743
No log 1.7 102 1.1009 0.4571 1.1009 1.0492
No log 1.7333 104 1.3157 0.5047 1.3157 1.1471
No log 1.7667 106 1.5555 0.4140 1.5555 1.2472
No log 1.8 108 1.6085 0.4016 1.6085 1.2683
No log 1.8333 110 1.4052 0.4621 1.4052 1.1854
No log 1.8667 112 1.0648 0.4965 1.0648 1.0319
No log 1.9 114 1.1494 0.4140 1.1494 1.0721
No log 1.9333 116 1.0893 0.4396 1.0893 1.0437
No log 1.9667 118 1.1368 0.4169 1.1368 1.0662
No log 2.0 120 1.2389 0.4124 1.2389 1.1131
No log 2.0333 122 1.3688 0.4076 1.3688 1.1700
No log 2.0667 124 1.0865 0.3845 1.0865 1.0423
No log 2.1 126 1.0282 0.4736 1.0282 1.0140
No log 2.1333 128 1.0783 0.4287 1.0783 1.0384
No log 2.1667 130 1.2546 0.4227 1.2546 1.1201
No log 2.2 132 1.2686 0.4800 1.2686 1.1263
No log 2.2333 134 1.3024 0.4694 1.3024 1.1412
No log 2.2667 136 1.1849 0.4448 1.1849 1.0885
No log 2.3 138 1.0570 0.4764 1.0570 1.0281
No log 2.3333 140 1.0845 0.5158 1.0845 1.0414
No log 2.3667 142 1.2552 0.4281 1.2552 1.1203
No log 2.4 144 1.1760 0.4823 1.1760 1.0845
No log 2.4333 146 1.0966 0.5027 1.0966 1.0472
No log 2.4667 148 1.1010 0.4954 1.1010 1.0493
No log 2.5 150 1.5945 0.3282 1.5945 1.2627
No log 2.5333 152 2.0389 0.2201 2.0389 1.4279
No log 2.5667 154 1.9399 0.2981 1.9399 1.3928
No log 2.6 156 1.3326 0.3927 1.3326 1.1544
No log 2.6333 158 1.1076 0.3974 1.1076 1.0524
No log 2.6667 160 1.0299 0.4524 1.0299 1.0149
No log 2.7 162 1.1710 0.4070 1.1710 1.0821
No log 2.7333 164 1.7128 0.3175 1.7128 1.3088
No log 2.7667 166 1.9868 0.2582 1.9868 1.4095
No log 2.8 168 1.5974 0.3194 1.5974 1.2639
No log 2.8333 170 1.0246 0.5549 1.0246 1.0122
No log 2.8667 172 0.9453 0.4604 0.9453 0.9723
No log 2.9 174 0.9265 0.4604 0.9265 0.9625
No log 2.9333 176 0.9339 0.4140 0.9339 0.9664
No log 2.9667 178 0.9732 0.4157 0.9732 0.9865
No log 3.0 180 1.0897 0.4371 1.0897 1.0439
No log 3.0333 182 1.0135 0.4913 1.0135 1.0067
No log 3.0667 184 0.9902 0.4790 0.9902 0.9951
No log 3.1 186 1.0005 0.4807 1.0005 1.0002
No log 3.1333 188 0.9105 0.5833 0.9105 0.9542
No log 3.1667 190 0.9025 0.5833 0.9025 0.9500
No log 3.2 192 1.0167 0.4978 1.0167 1.0083
No log 3.2333 194 1.0415 0.4571 1.0415 1.0205
No log 3.2667 196 0.9506 0.4718 0.9506 0.9750
No log 3.3 198 0.9124 0.5049 0.9124 0.9552
No log 3.3333 200 0.9562 0.4925 0.9562 0.9779
No log 3.3667 202 0.9814 0.5236 0.9814 0.9906
No log 3.4 204 1.0255 0.4681 1.0255 1.0127
No log 3.4333 206 1.0164 0.4708 1.0164 1.0082
No log 3.4667 208 0.8923 0.5779 0.8923 0.9446
No log 3.5 210 0.8854 0.5645 0.8854 0.9410
No log 3.5333 212 0.9681 0.5392 0.9681 0.9839
No log 3.5667 214 0.9708 0.5133 0.9708 0.9853
No log 3.6 216 0.9212 0.6355 0.9212 0.9598
No log 3.6333 218 0.9394 0.5816 0.9394 0.9692
No log 3.6667 220 1.0976 0.4550 1.0976 1.0477
No log 3.7 222 1.0687 0.5028 1.0687 1.0338
No log 3.7333 224 0.9447 0.5744 0.9447 0.9720
No log 3.7667 226 0.9031 0.5426 0.9031 0.9503
No log 3.8 228 0.9534 0.5816 0.9534 0.9764
No log 3.8333 230 1.0242 0.5303 1.0242 1.0120
No log 3.8667 232 1.1183 0.4403 1.1183 1.0575
No log 3.9 234 0.9965 0.4139 0.9965 0.9982
No log 3.9333 236 0.9162 0.4288 0.9162 0.9572
No log 3.9667 238 0.9285 0.4196 0.9285 0.9636
No log 4.0 240 1.0353 0.3727 1.0353 1.0175
No log 4.0333 242 1.4974 0.3581 1.4974 1.2237
No log 4.0667 244 1.7846 0.2891 1.7846 1.3359
No log 4.1 246 1.6631 0.3105 1.6631 1.2896
No log 4.1333 248 1.3295 0.3311 1.3295 1.1531
No log 4.1667 250 1.0748 0.3330 1.0748 1.0367
No log 4.2 252 1.0443 0.3506 1.0443 1.0219
No log 4.2333 254 0.9718 0.2967 0.9718 0.9858
No log 4.2667 256 0.9544 0.4196 0.9544 0.9769
No log 4.3 258 0.9544 0.4818 0.9544 0.9769
No log 4.3333 260 0.9887 0.4681 0.9887 0.9944
No log 4.3667 262 0.9929 0.5008 0.9929 0.9965
No log 4.4 264 0.9441 0.4847 0.9441 0.9717
No log 4.4333 266 0.9305 0.5295 0.9305 0.9646
No log 4.4667 268 0.9243 0.5114 0.9243 0.9614
No log 4.5 270 0.9429 0.3871 0.9429 0.9710
No log 4.5333 272 0.9437 0.4050 0.9437 0.9714
No log 4.5667 274 0.9589 0.4121 0.9589 0.9792
No log 4.6 276 0.9392 0.4017 0.9392 0.9691
No log 4.6333 278 0.9372 0.4196 0.9372 0.9681
No log 4.6667 280 0.9372 0.4435 0.9372 0.9681
No log 4.7 282 0.9547 0.5037 0.9547 0.9771
No log 4.7333 284 1.0586 0.4544 1.0586 1.0289
No log 4.7667 286 1.0599 0.4627 1.0599 1.0295
No log 4.8 288 1.0549 0.4882 1.0549 1.0271
No log 4.8333 290 1.0797 0.4673 1.0797 1.0391
No log 4.8667 292 1.1788 0.4431 1.1788 1.0857
No log 4.9 294 1.4138 0.3056 1.4138 1.1890
No log 4.9333 296 1.3110 0.3129 1.3110 1.1450
No log 4.9667 298 1.0435 0.3146 1.0435 1.0215
No log 5.0 300 1.0036 0.2967 1.0036 1.0018
No log 5.0333 302 1.0319 0.3433 1.0319 1.0158
No log 5.0667 304 1.1391 0.3923 1.1391 1.0673
No log 5.1 306 1.1402 0.3612 1.1402 1.0678
No log 5.1333 308 1.0507 0.3213 1.0507 1.0250
No log 5.1667 310 0.9974 0.3542 0.9974 0.9987
No log 5.2 312 1.0413 0.3149 1.0413 1.0205
No log 5.2333 314 1.0688 0.3008 1.0688 1.0338
No log 5.2667 316 1.0609 0.3622 1.0609 1.0300
No log 5.3 318 0.9699 0.3262 0.9699 0.9849
No log 5.3333 320 0.9073 0.4388 0.9073 0.9525
No log 5.3667 322 0.8959 0.5042 0.8959 0.9465
No log 5.4 324 0.9034 0.5178 0.9034 0.9505
No log 5.4333 326 0.9967 0.3768 0.9967 0.9984
No log 5.4667 328 1.1884 0.3820 1.1884 1.0902
No log 5.5 330 1.1824 0.3989 1.1824 1.0874
No log 5.5333 332 0.9631 0.3856 0.9631 0.9814
No log 5.5667 334 0.8873 0.5251 0.8873 0.9420
No log 5.6 336 0.8946 0.5251 0.8946 0.9458
No log 5.6333 338 0.9794 0.3933 0.9794 0.9897
No log 5.6667 340 1.1675 0.3907 1.1675 1.0805
No log 5.7 342 1.2313 0.3825 1.2313 1.1096
No log 5.7333 344 1.1509 0.3902 1.1509 1.0728
No log 5.7667 346 0.9548 0.4018 0.9548 0.9772
No log 5.8 348 0.9179 0.4516 0.9179 0.9581
No log 5.8333 350 1.0055 0.3763 1.0055 1.0027
No log 5.8667 352 1.0258 0.4086 1.0258 1.0128
No log 5.9 354 0.9993 0.3841 0.9993 0.9996
No log 5.9333 356 1.0466 0.4086 1.0466 1.0230
No log 5.9667 358 0.9741 0.3546 0.9741 0.9870
No log 6.0 360 0.9391 0.3378 0.9391 0.9691
No log 6.0333 362 1.0045 0.3719 1.0045 1.0022
No log 6.0667 364 1.1213 0.4154 1.1213 1.0589
No log 6.1 366 1.1227 0.4120 1.1227 1.0596
No log 6.1333 368 0.9466 0.3243 0.9466 0.9729
No log 6.1667 370 0.9046 0.5076 0.9046 0.9511
No log 6.2 372 0.9596 0.4394 0.9596 0.9796
No log 6.2333 374 0.9345 0.4575 0.9345 0.9667
No log 6.2667 376 0.9145 0.3979 0.9145 0.9563
No log 6.3 378 0.9546 0.3439 0.9546 0.9771
No log 6.3333 380 0.9659 0.3439 0.9659 0.9828
No log 6.3667 382 0.9179 0.4771 0.9179 0.9581
No log 6.4 384 0.9272 0.4690 0.9272 0.9629
No log 6.4333 386 0.9294 0.5045 0.9294 0.9641
No log 6.4667 388 0.9274 0.4120 0.9274 0.9630
No log 6.5 390 0.9503 0.3775 0.9503 0.9748
No log 6.5333 392 1.0374 0.3877 1.0374 1.0185
No log 6.5667 394 1.1420 0.3750 1.1420 1.0687
No log 6.6 396 1.0435 0.4086 1.0435 1.0215
No log 6.6333 398 0.9293 0.5056 0.9293 0.9640
No log 6.6667 400 0.9559 0.4236 0.9559 0.9777
No log 6.7 402 0.9462 0.4270 0.9462 0.9727
No log 6.7333 404 0.9367 0.4778 0.9367 0.9678
No log 6.7667 406 0.9472 0.3379 0.9472 0.9732
No log 6.8 408 1.0057 0.3976 1.0057 1.0028
No log 6.8333 410 1.0440 0.3963 1.0440 1.0218
No log 6.8667 412 1.0544 0.4237 1.0544 1.0268
No log 6.9 414 1.0948 0.4247 1.0948 1.0463
No log 6.9333 416 0.9997 0.4612 0.9997 0.9998
No log 6.9667 418 0.8638 0.4388 0.8638 0.9294
No log 7.0 420 0.8695 0.5076 0.8695 0.9325
No log 7.0333 422 0.8554 0.5076 0.8554 0.9249
No log 7.0667 424 0.8706 0.4361 0.8706 0.9331
No log 7.1 426 0.9598 0.3850 0.9598 0.9797
No log 7.1333 428 0.9435 0.3812 0.9435 0.9713
No log 7.1667 430 0.8963 0.4084 0.8963 0.9467
No log 7.2 432 0.8851 0.4392 0.8851 0.9408
No log 7.2333 434 0.9829 0.4172 0.9829 0.9914
No log 7.2667 436 1.0490 0.4155 1.0490 1.0242
No log 7.3 438 0.9965 0.3805 0.9965 0.9982
No log 7.3333 440 0.9018 0.4548 0.9018 0.9496
No log 7.3667 442 0.8921 0.5167 0.8921 0.9445
No log 7.4 444 0.9210 0.4102 0.9210 0.9597
No log 7.4333 446 0.9915 0.3213 0.9915 0.9957
No log 7.4667 448 0.9949 0.3078 0.9949 0.9974
No log 7.5 450 0.9997 0.2714 0.9997 0.9999
No log 7.5333 452 1.0269 0.2896 1.0269 1.0133
No log 7.5667 454 1.0100 0.3099 1.0100 1.0050
No log 7.6 456 0.9840 0.3775 0.9840 0.9919
No log 7.6333 458 0.9236 0.4749 0.9236 0.9610
No log 7.6667 460 0.9197 0.5267 0.9197 0.9590
No log 7.7 462 0.9263 0.5026 0.9263 0.9624
No log 7.7333 464 0.9533 0.4139 0.9533 0.9764
No log 7.7667 466 0.9526 0.4159 0.9526 0.9760
No log 7.8 468 0.9688 0.3829 0.9688 0.9843
No log 7.8333 470 0.9712 0.3829 0.9712 0.9855
No log 7.8667 472 0.9819 0.3312 0.9819 0.9909
No log 7.9 474 1.0299 0.3439 1.0299 1.0148
No log 7.9333 476 1.0191 0.3714 1.0191 1.0095
No log 7.9667 478 1.0286 0.3404 1.0286 1.0142
No log 8.0 480 0.9840 0.3009 0.9840 0.9920
No log 8.0333 482 0.9858 0.3363 0.9858 0.9929
No log 8.0667 484 0.9799 0.3126 0.9799 0.9899
No log 8.1 486 0.9775 0.3218 0.9775 0.9887
No log 8.1333 488 0.9675 0.3733 0.9675 0.9836
No log 8.1667 490 0.9724 0.2896 0.9724 0.9861
No log 8.2 492 0.9616 0.3218 0.9616 0.9806
No log 8.2333 494 0.9398 0.4420 0.9398 0.9694
No log 8.2667 496 0.9266 0.4420 0.9266 0.9626
No log 8.3 498 0.9320 0.4326 0.9320 0.9654
0.3458 8.3333 500 0.9388 0.3596 0.9388 0.9689
0.3458 8.3667 502 0.9433 0.3554 0.9433 0.9712
0.3458 8.4 504 0.9642 0.3013 0.9642 0.9819
0.3458 8.4333 506 1.0137 0.2984 1.0137 1.0068
0.3458 8.4667 508 1.0097 0.2984 1.0097 1.0049
0.3458 8.5 510 1.0057 0.2984 1.0057 1.0029

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k12_task2_organization

Finetuned
(4019)
this model