ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k13_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9285
  • Qwk: 0.3403
  • Mse: 0.9285
  • Rmse: 0.9636

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0328 2 4.1553 -0.0139 4.1553 2.0385
No log 0.0656 4 2.3328 -0.0086 2.3328 1.5274
No log 0.0984 6 1.8644 -0.0307 1.8644 1.3654
No log 0.1311 8 1.2243 0.1468 1.2243 1.1065
No log 0.1639 10 1.1154 0.0947 1.1154 1.0561
No log 0.1967 12 1.1375 0.1453 1.1375 1.0665
No log 0.2295 14 1.3608 0.1533 1.3608 1.1665
No log 0.2623 16 1.1768 0.3160 1.1768 1.0848
No log 0.2951 18 0.9992 0.2449 0.9992 0.9996
No log 0.3279 20 1.0202 0.2049 1.0202 1.0101
No log 0.3607 22 1.3808 0.2103 1.3808 1.1751
No log 0.3934 24 1.4477 0.2328 1.4477 1.2032
No log 0.4262 26 1.0381 0.3043 1.0381 1.0189
No log 0.4590 28 1.1311 0.3409 1.1311 1.0635
No log 0.4918 30 1.0563 0.3349 1.0563 1.0278
No log 0.5246 32 1.0442 0.3646 1.0442 1.0219
No log 0.5574 34 1.2922 0.2087 1.2922 1.1368
No log 0.5902 36 1.4417 0.1979 1.4417 1.2007
No log 0.6230 38 1.1422 0.2954 1.1422 1.0687
No log 0.6557 40 1.0008 0.3428 1.0008 1.0004
No log 0.6885 42 1.0120 0.3059 1.0120 1.0060
No log 0.7213 44 1.1175 0.3090 1.1175 1.0571
No log 0.7541 46 1.3225 0.1793 1.3225 1.1500
No log 0.7869 48 1.2976 0.2149 1.2976 1.1391
No log 0.8197 50 1.0504 0.1727 1.0504 1.0249
No log 0.8525 52 1.0008 0.3689 1.0008 1.0004
No log 0.8852 54 1.0163 0.2647 1.0163 1.0081
No log 0.9180 56 1.3200 0.2881 1.3200 1.1489
No log 0.9508 58 1.5033 0.2292 1.5033 1.2261
No log 0.9836 60 1.4751 0.2292 1.4751 1.2146
No log 1.0164 62 1.5465 0.2042 1.5465 1.2436
No log 1.0492 64 1.0733 0.3243 1.0733 1.0360
No log 1.0820 66 1.0053 0.4213 1.0053 1.0027
No log 1.1148 68 1.0353 0.3853 1.0353 1.0175
No log 1.1475 70 1.2800 0.3184 1.2800 1.1314
No log 1.1803 72 1.3943 0.3080 1.3943 1.1808
No log 1.2131 74 1.3921 0.2942 1.3921 1.1799
No log 1.2459 76 1.3622 0.2834 1.3622 1.1671
No log 1.2787 78 1.1843 0.3590 1.1843 1.0882
No log 1.3115 80 1.0891 0.3727 1.0891 1.0436
No log 1.3443 82 1.0840 0.3821 1.0840 1.0412
No log 1.3770 84 1.3635 0.2779 1.3635 1.1677
No log 1.4098 86 1.7373 0.1809 1.7373 1.3180
No log 1.4426 88 1.5365 0.2336 1.5365 1.2395
No log 1.4754 90 1.1270 0.3208 1.1270 1.0616
No log 1.5082 92 0.9861 0.4039 0.9861 0.9930
No log 1.5410 94 0.9983 0.3804 0.9983 0.9992
No log 1.5738 96 1.1286 0.2626 1.1286 1.0623
No log 1.6066 98 1.2527 0.2240 1.2527 1.1192
No log 1.6393 100 1.0867 0.2693 1.0867 1.0424
No log 1.6721 102 1.1662 0.2968 1.1662 1.0799
No log 1.7049 104 1.2881 0.2516 1.2881 1.1349
No log 1.7377 106 1.1784 0.3546 1.1784 1.0855
No log 1.7705 108 1.0120 0.3485 1.0120 1.0060
No log 1.8033 110 1.1383 0.1318 1.1383 1.0669
No log 1.8361 112 1.2068 0.2574 1.2068 1.0986
No log 1.8689 114 1.0570 0.2625 1.0570 1.0281
No log 1.9016 116 1.0454 0.3590 1.0454 1.0224
No log 1.9344 118 1.1787 0.3495 1.1787 1.0857
No log 1.9672 120 1.1331 0.3945 1.1331 1.0645
No log 2.0 122 1.0885 0.3343 1.0885 1.0433
No log 2.0328 124 1.5090 0.1594 1.5090 1.2284
No log 2.0656 126 1.6693 0.1996 1.6693 1.2920
No log 2.0984 128 1.3824 0.1187 1.3824 1.1758
No log 2.1311 130 1.0880 0.3437 1.0880 1.0431
No log 2.1639 132 1.1133 0.3250 1.1133 1.0551
No log 2.1967 134 1.0752 0.3145 1.0752 1.0369
No log 2.2295 136 1.0472 0.3154 1.0472 1.0233
No log 2.2623 138 1.1969 0.1943 1.1969 1.0940
No log 2.2951 140 1.3530 0.0990 1.3530 1.1632
No log 2.3279 142 1.1894 0.1807 1.1894 1.0906
No log 2.3607 144 1.0513 0.3378 1.0513 1.0253
No log 2.3934 146 1.1065 0.3065 1.1065 1.0519
No log 2.4262 148 1.0782 0.3243 1.0782 1.0384
No log 2.4590 150 1.0737 0.3950 1.0737 1.0362
No log 2.4918 152 1.0714 0.3559 1.0714 1.0351
No log 2.5246 154 1.1277 0.3250 1.1277 1.0619
No log 2.5574 156 1.1145 0.3250 1.1145 1.0557
No log 2.5902 158 1.0914 0.3753 1.0914 1.0447
No log 2.6230 160 1.0614 0.3844 1.0614 1.0302
No log 2.6557 162 0.9690 0.3455 0.9690 0.9844
No log 2.6885 164 0.9479 0.3949 0.9479 0.9736
No log 2.7213 166 0.9560 0.4197 0.9560 0.9777
No log 2.7541 168 0.9501 0.3785 0.9501 0.9747
No log 2.7869 170 0.9360 0.3983 0.9360 0.9675
No log 2.8197 172 0.9478 0.4164 0.9478 0.9736
No log 2.8525 174 0.9421 0.3806 0.9421 0.9706
No log 2.8852 176 0.9132 0.3666 0.9132 0.9556
No log 2.9180 178 0.9418 0.3680 0.9418 0.9705
No log 2.9508 180 0.9823 0.3834 0.9823 0.9911
No log 2.9836 182 0.9298 0.4337 0.9298 0.9643
No log 3.0164 184 1.0105 0.4275 1.0105 1.0053
No log 3.0492 186 1.0200 0.4283 1.0200 1.0100
No log 3.0820 188 0.9575 0.4555 0.9575 0.9785
No log 3.1148 190 0.9789 0.4337 0.9789 0.9894
No log 3.1475 192 1.1757 0.2672 1.1757 1.0843
No log 3.1803 194 1.2728 0.2405 1.2728 1.1282
No log 3.2131 196 1.0488 0.3068 1.0488 1.0241
No log 3.2459 198 0.9174 0.3493 0.9174 0.9578
No log 3.2787 200 0.9990 0.3753 0.9990 0.9995
No log 3.3115 202 1.0071 0.4020 1.0071 1.0035
No log 3.3443 204 0.9718 0.4381 0.9718 0.9858
No log 3.3770 206 1.0822 0.4007 1.0822 1.0403
No log 3.4098 208 1.0591 0.3284 1.0591 1.0291
No log 3.4426 210 0.9626 0.4381 0.9626 0.9811
No log 3.4754 212 0.9809 0.4042 0.9809 0.9904
No log 3.5082 214 1.0235 0.4220 1.0235 1.0117
No log 3.5410 216 0.9959 0.3725 0.9959 0.9980
No log 3.5738 218 0.9680 0.4787 0.9680 0.9839
No log 3.6066 220 0.9738 0.4676 0.9738 0.9868
No log 3.6393 222 1.0100 0.3338 1.0100 1.0050
No log 3.6721 224 1.0722 0.3062 1.0722 1.0355
No log 3.7049 226 1.0161 0.2723 1.0161 1.0080
No log 3.7377 228 0.9355 0.3301 0.9355 0.9672
No log 3.7705 230 0.9491 0.3757 0.9491 0.9742
No log 3.8033 232 0.9547 0.3757 0.9547 0.9771
No log 3.8361 234 0.9542 0.3781 0.9542 0.9768
No log 3.8689 236 0.9681 0.3652 0.9681 0.9839
No log 3.9016 238 1.0136 0.3300 1.0136 1.0068
No log 3.9344 240 0.9923 0.3446 0.9923 0.9961
No log 3.9672 242 0.9984 0.3554 0.9984 0.9992
No log 4.0 244 1.0213 0.3463 1.0213 1.0106
No log 4.0328 246 0.9575 0.2480 0.9575 0.9785
No log 4.0656 248 0.9495 0.2624 0.9495 0.9744
No log 4.0984 250 1.0081 0.2200 1.0081 1.0040
No log 4.1311 252 1.0686 0.2155 1.0686 1.0337
No log 4.1639 254 1.0509 0.2424 1.0509 1.0252
No log 4.1967 256 1.0467 0.3284 1.0467 1.0231
No log 4.2295 258 1.0461 0.2696 1.0461 1.0228
No log 4.2623 260 1.0699 0.2723 1.0699 1.0343
No log 4.2951 262 1.0634 0.2702 1.0634 1.0312
No log 4.3279 264 0.9975 0.2746 0.9975 0.9987
No log 4.3607 266 0.9368 0.4288 0.9368 0.9679
No log 4.3934 268 0.9442 0.3873 0.9442 0.9717
No log 4.4262 270 0.9336 0.4337 0.9336 0.9662
No log 4.4590 272 0.9834 0.3437 0.9834 0.9917
No log 4.4918 274 1.0304 0.2787 1.0304 1.0151
No log 4.5246 276 1.0037 0.3437 1.0037 1.0018
No log 4.5574 278 0.9724 0.3184 0.9724 0.9861
No log 4.5902 280 0.9589 0.4100 0.9589 0.9792
No log 4.6230 282 0.9539 0.3986 0.9539 0.9767
No log 4.6557 284 0.9200 0.3760 0.9200 0.9591
No log 4.6885 286 0.9289 0.3243 0.9289 0.9638
No log 4.7213 288 0.9280 0.2909 0.9280 0.9633
No log 4.7541 290 0.9017 0.3347 0.9017 0.9496
No log 4.7869 292 0.8952 0.3616 0.8952 0.9461
No log 4.8197 294 0.9981 0.4315 0.9981 0.9991
No log 4.8525 296 0.9627 0.4769 0.9627 0.9812
No log 4.8852 298 0.8661 0.3476 0.8661 0.9307
No log 4.9180 300 0.9403 0.2698 0.9403 0.9697
No log 4.9508 302 0.9835 0.3161 0.9835 0.9917
No log 4.9836 304 0.9514 0.3634 0.9514 0.9754
No log 5.0164 306 0.9751 0.3634 0.9751 0.9875
No log 5.0492 308 1.0155 0.3119 1.0155 1.0077
No log 5.0820 310 0.9614 0.3326 0.9614 0.9805
No log 5.1148 312 0.9278 0.3577 0.9278 0.9632
No log 5.1475 314 0.9398 0.3685 0.9398 0.9694
No log 5.1803 316 0.9315 0.3030 0.9315 0.9651
No log 5.2131 318 0.9393 0.3030 0.9393 0.9692
No log 5.2459 320 0.9486 0.2818 0.9486 0.9740
No log 5.2787 322 0.9456 0.2643 0.9456 0.9724
No log 5.3115 324 0.9468 0.2492 0.9468 0.9731
No log 5.3443 326 0.9639 0.2818 0.9639 0.9818
No log 5.3770 328 0.9626 0.3030 0.9626 0.9811
No log 5.4098 330 0.9915 0.3967 0.9915 0.9957
No log 5.4426 332 1.0062 0.4017 1.0062 1.0031
No log 5.4754 334 0.9669 0.4014 0.9669 0.9833
No log 5.5082 336 1.0525 0.2878 1.0525 1.0259
No log 5.5410 338 1.1092 0.3099 1.1092 1.0532
No log 5.5738 340 1.0337 0.2949 1.0337 1.0167
No log 5.6066 342 0.9490 0.2667 0.9490 0.9742
No log 5.6393 344 1.0072 0.3676 1.0072 1.0036
No log 5.6721 346 0.9937 0.4338 0.9937 0.9969
No log 5.7049 348 0.9324 0.2667 0.9324 0.9656
No log 5.7377 350 0.9530 0.3005 0.9530 0.9762
No log 5.7705 352 1.0287 0.3424 1.0287 1.0143
No log 5.8033 354 1.0019 0.3062 1.0019 1.0010
No log 5.8361 356 0.9388 0.3483 0.9388 0.9689
No log 5.8689 358 0.8765 0.3896 0.8765 0.9362
No log 5.9016 360 0.8736 0.3896 0.8736 0.9346
No log 5.9344 362 0.8970 0.3675 0.8970 0.9471
No log 5.9672 364 1.0237 0.3845 1.0237 1.0118
No log 6.0 366 1.1770 0.3578 1.1770 1.0849
No log 6.0328 368 1.1706 0.3578 1.1706 1.0819
No log 6.0656 370 1.0443 0.3434 1.0443 1.0219
No log 6.0984 372 0.9235 0.3338 0.9235 0.9610
No log 6.1311 374 0.9093 0.3493 0.9093 0.9536
No log 6.1639 376 0.9600 0.3483 0.9600 0.9798
No log 6.1967 378 1.0323 0.3862 1.0323 1.0160
No log 6.2295 380 1.0408 0.3499 1.0408 1.0202
No log 6.2623 382 0.9708 0.3237 0.9708 0.9853
No log 6.2951 384 0.9362 0.3609 0.9362 0.9676
No log 6.3279 386 0.9368 0.3182 0.9368 0.9679
No log 6.3607 388 0.9654 0.2582 0.9654 0.9826
No log 6.3934 390 0.9373 0.2263 0.9373 0.9682
No log 6.4262 392 0.9327 0.1573 0.9327 0.9658
No log 6.4590 394 0.9369 0.1573 0.9369 0.9680
No log 6.4918 396 0.9181 0.1881 0.9181 0.9582
No log 6.5246 398 0.9150 0.2647 0.9150 0.9565
No log 6.5574 400 0.9438 0.2582 0.9438 0.9715
No log 6.5902 402 0.9741 0.2886 0.9741 0.9870
No log 6.6230 404 0.9402 0.2836 0.9402 0.9696
No log 6.6557 406 0.9198 0.3250 0.9198 0.9591
No log 6.6885 408 0.8636 0.4073 0.8636 0.9293
No log 6.7213 410 0.8699 0.4361 0.8699 0.9327
No log 6.7541 412 0.8927 0.4361 0.8927 0.9448
No log 6.7869 414 0.8799 0.3839 0.8799 0.9380
No log 6.8197 416 0.8962 0.3347 0.8962 0.9467
No log 6.8525 418 0.9056 0.3133 0.9056 0.9516
No log 6.8852 420 0.9307 0.3704 0.9307 0.9647
No log 6.9180 422 0.9244 0.2932 0.9244 0.9614
No log 6.9508 424 0.9248 0.2596 0.9248 0.9617
No log 6.9836 426 0.9186 0.3284 0.9186 0.9584
No log 7.0164 428 0.9214 0.3418 0.9214 0.9599
No log 7.0492 430 0.9659 0.3985 0.9659 0.9828
No log 7.0820 432 1.0203 0.3415 1.0203 1.0101
No log 7.1148 434 0.9639 0.4130 0.9639 0.9818
No log 7.1475 436 0.9292 0.4010 0.9292 0.9640
No log 7.1803 438 0.9887 0.3551 0.9887 0.9944
No log 7.2131 440 0.9690 0.3706 0.9690 0.9844
No log 7.2459 442 0.9270 0.3151 0.9270 0.9628
No log 7.2787 444 0.9179 0.2517 0.9179 0.9581
No log 7.3115 446 0.9178 0.2932 0.9178 0.9580
No log 7.3443 448 0.9221 0.3229 0.9221 0.9603
No log 7.3770 450 0.9747 0.3822 0.9747 0.9873
No log 7.4098 452 1.0197 0.3957 1.0197 1.0098
No log 7.4426 454 0.9794 0.3087 0.9794 0.9896
No log 7.4754 456 0.9371 0.3744 0.9371 0.9680
No log 7.5082 458 0.9191 0.3744 0.9191 0.9587
No log 7.5410 460 0.9317 0.3202 0.9317 0.9652
No log 7.5738 462 0.9573 0.3577 0.9573 0.9784
No log 7.6066 464 0.9611 0.4111 0.9611 0.9804
No log 7.6393 466 0.9396 0.4111 0.9396 0.9693
No log 7.6721 468 0.8753 0.4264 0.8753 0.9356
No log 7.7049 470 0.8782 0.4391 0.8782 0.9371
No log 7.7377 472 0.8845 0.3879 0.8845 0.9405
No log 7.7705 474 0.9396 0.3330 0.9396 0.9694
No log 7.8033 476 0.9859 0.3987 0.9859 0.9929
No log 7.8361 478 0.9528 0.4111 0.9528 0.9761
No log 7.8689 480 0.9313 0.4004 0.9313 0.9650
No log 7.9016 482 0.8896 0.3976 0.8896 0.9432
No log 7.9344 484 0.8903 0.3976 0.8903 0.9436
No log 7.9672 486 0.9268 0.4352 0.9268 0.9627
No log 8.0 488 1.0090 0.4342 1.0090 1.0045
No log 8.0328 490 1.0867 0.3191 1.0867 1.0424
No log 8.0656 492 1.0677 0.2179 1.0677 1.0333
No log 8.0984 494 0.9879 0.3145 0.9879 0.9939
No log 8.1311 496 0.9916 0.3354 0.9916 0.9958
No log 8.1639 498 1.0300 0.3098 1.0300 1.0149
0.2967 8.1967 500 1.0102 0.3440 1.0102 1.0051
0.2967 8.2295 502 0.9632 0.3457 0.9632 0.9814
0.2967 8.2623 504 0.9736 0.3706 0.9736 0.9867
0.2967 8.2951 506 1.0622 0.3699 1.0622 1.0306
0.2967 8.3279 508 1.0470 0.3434 1.0470 1.0232
0.2967 8.3607 510 0.9557 0.2723 0.9557 0.9776
0.2967 8.3934 512 0.9147 0.2643 0.9147 0.9564
0.2967 8.4262 514 0.9270 0.3425 0.9270 0.9628
0.2967 8.4590 516 0.9285 0.3403 0.9285 0.9636

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k13_task5_organization

Finetuned
(4019)
this model