ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k7_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9704
  • Qwk: 0.6494
  • Mse: 0.9704
  • Rmse: 0.9851

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0625 2 7.4937 -0.0587 7.4937 2.7375
No log 0.125 4 4.4141 0.0675 4.4141 2.1010
No log 0.1875 6 3.2175 0.0351 3.2175 1.7937
No log 0.25 8 2.6869 0.1119 2.6869 1.6392
No log 0.3125 10 2.1050 0.1148 2.1050 1.4509
No log 0.375 12 1.7689 0.1495 1.7689 1.3300
No log 0.4375 14 1.7853 0.0556 1.7853 1.3362
No log 0.5 16 2.0506 -0.0500 2.0506 1.4320
No log 0.5625 18 2.9536 0.0519 2.9536 1.7186
No log 0.625 20 4.1244 0.0180 4.1244 2.0309
No log 0.6875 22 3.6235 0.0821 3.6235 1.9036
No log 0.75 24 2.4066 0.1589 2.4066 1.5513
No log 0.8125 26 1.8751 0.3556 1.8751 1.3693
No log 0.875 28 1.9209 0.25 1.9209 1.3860
No log 0.9375 30 2.2884 0.0146 2.2884 1.5127
No log 1.0 32 2.4472 -0.0284 2.4472 1.5643
No log 1.0625 34 2.3431 -0.0141 2.3431 1.5307
No log 1.125 36 1.9615 0.2903 1.9615 1.4005
No log 1.1875 38 1.7912 0.2833 1.7912 1.3384
No log 1.25 40 1.8816 0.2812 1.8816 1.3717
No log 1.3125 42 2.0919 0.2941 2.0919 1.4463
No log 1.375 44 2.0123 0.2628 2.0123 1.4185
No log 1.4375 46 1.7844 0.3664 1.7844 1.3358
No log 1.5 48 1.7583 0.3485 1.7583 1.3260
No log 1.5625 50 2.0073 0.2628 2.0073 1.4168
No log 1.625 52 2.2164 0.2000 2.2164 1.4887
No log 1.6875 54 2.1623 0.2158 2.1623 1.4705
No log 1.75 56 2.1670 0.2113 2.1670 1.4721
No log 1.8125 58 2.4966 0.1410 2.4966 1.5801
No log 1.875 60 3.2916 0.1981 3.2916 1.8143
No log 1.9375 62 3.5404 0.2241 3.5404 1.8816
No log 2.0 64 2.8679 0.1622 2.8679 1.6935
No log 2.0625 66 2.4088 0.2038 2.4088 1.5520
No log 2.125 68 1.8660 0.3333 1.8660 1.3660
No log 2.1875 70 1.7434 0.2927 1.7434 1.3204
No log 2.25 72 1.7749 0.3281 1.7749 1.3323
No log 2.3125 74 2.0875 0.1884 2.0875 1.4448
No log 2.375 76 2.6296 0.125 2.6296 1.6216
No log 2.4375 78 3.3215 0.1263 3.3215 1.8225
No log 2.5 80 4.1444 0.1285 4.1444 2.0358
No log 2.5625 82 3.5114 0.2081 3.5114 1.8739
No log 2.625 84 2.5785 0.2727 2.5785 1.6058
No log 2.6875 86 2.2309 0.2840 2.2309 1.4936
No log 2.75 88 1.9141 0.2676 1.9141 1.3835
No log 2.8125 90 2.0339 0.2238 2.0339 1.4261
No log 2.875 92 2.0444 0.2632 2.0444 1.4298
No log 2.9375 94 2.0338 0.2981 2.0338 1.4261
No log 3.0 96 1.9611 0.3509 1.9611 1.4004
No log 3.0625 98 2.2477 0.4098 2.2477 1.4992
No log 3.125 100 2.3606 0.4112 2.3606 1.5364
No log 3.1875 102 2.0836 0.4021 2.0836 1.4435
No log 3.25 104 1.8831 0.3291 1.8831 1.3723
No log 3.3125 106 1.5597 0.3407 1.5597 1.2489
No log 3.375 108 1.4129 0.4065 1.4129 1.1886
No log 3.4375 110 1.4215 0.4065 1.4215 1.1923
No log 3.5 112 1.5227 0.375 1.5227 1.2340
No log 3.5625 114 1.5774 0.3913 1.5774 1.2560
No log 3.625 116 1.5291 0.4539 1.5291 1.2366
No log 3.6875 118 1.6541 0.4459 1.6541 1.2861
No log 3.75 120 1.9363 0.3882 1.9363 1.3915
No log 3.8125 122 2.1934 0.4444 2.1934 1.4810
No log 3.875 124 2.6329 0.3519 2.6329 1.6226
No log 3.9375 126 2.7509 0.3349 2.7509 1.6586
No log 4.0 128 2.7788 0.3349 2.7788 1.6670
No log 4.0625 130 2.6747 0.3333 2.6747 1.6355
No log 4.125 132 2.3598 0.3316 2.3598 1.5362
No log 4.1875 134 2.0915 0.4 2.0915 1.4462
No log 4.25 136 1.6268 0.4 1.6268 1.2755
No log 4.3125 138 1.6592 0.3974 1.6592 1.2881
No log 4.375 140 1.8957 0.3226 1.8957 1.3769
No log 4.4375 142 1.9338 0.3396 1.9338 1.3906
No log 4.5 144 1.8866 0.4118 1.8866 1.3735
No log 4.5625 146 1.5709 0.4512 1.5709 1.2533
No log 4.625 148 1.3225 0.5385 1.3225 1.1500
No log 4.6875 150 1.1365 0.5946 1.1365 1.0661
No log 4.75 152 1.1447 0.6 1.1447 1.0699
No log 4.8125 154 1.1803 0.6093 1.1803 1.0864
No log 4.875 156 1.5133 0.4938 1.5133 1.2302
No log 4.9375 158 1.8054 0.4253 1.8054 1.3437
No log 5.0 160 1.9947 0.4468 1.9947 1.4123
No log 5.0625 162 1.9840 0.4398 1.9840 1.4085
No log 5.125 164 1.8932 0.4762 1.8932 1.3759
No log 5.1875 166 1.5051 0.4938 1.5051 1.2268
No log 5.25 168 1.2284 0.5621 1.2284 1.1083
No log 5.3125 170 1.0523 0.6099 1.0523 1.0258
No log 5.375 172 1.0422 0.5882 1.0422 1.0209
No log 5.4375 174 1.0829 0.5797 1.0829 1.0406
No log 5.5 176 1.3578 0.5556 1.3578 1.1652
No log 5.5625 178 1.9939 0.4821 1.9939 1.4121
No log 5.625 180 2.3717 0.4444 2.3717 1.5400
No log 5.6875 182 2.2367 0.4532 2.2367 1.4955
No log 5.75 184 1.8170 0.5279 1.8170 1.3480
No log 5.8125 186 1.7643 0.5155 1.7643 1.3283
No log 5.875 188 1.8927 0.5178 1.8927 1.3757
No log 5.9375 190 1.8506 0.5128 1.8506 1.3604
No log 6.0 192 1.7222 0.5 1.7222 1.3123
No log 6.0625 194 1.5622 0.4971 1.5622 1.2499
No log 6.125 196 1.3707 0.5096 1.3707 1.1708
No log 6.1875 198 1.3126 0.4967 1.3126 1.1457
No log 6.25 200 1.2910 0.5065 1.2910 1.1362
No log 6.3125 202 1.2964 0.4837 1.2964 1.1386
No log 6.375 204 1.4051 0.4875 1.4051 1.1854
No log 6.4375 206 1.4443 0.4720 1.4443 1.2018
No log 6.5 208 1.4123 0.5060 1.4123 1.1884
No log 6.5625 210 1.2793 0.5409 1.2793 1.1311
No log 6.625 212 1.2920 0.5556 1.2920 1.1367
No log 6.6875 214 1.4725 0.5172 1.4725 1.2135
No log 6.75 216 1.3794 0.5517 1.3794 1.1745
No log 6.8125 218 1.2461 0.6190 1.2461 1.1163
No log 6.875 220 1.0863 0.5660 1.0863 1.0423
No log 6.9375 222 0.9495 0.6410 0.9495 0.9744
No log 7.0 224 0.9306 0.675 0.9306 0.9647
No log 7.0625 226 0.9601 0.6289 0.9601 0.9799
No log 7.125 228 1.1070 0.5904 1.1070 1.0521
No log 7.1875 230 1.3024 0.5922 1.3024 1.1412
No log 7.25 232 1.2849 0.6 1.2849 1.1335
No log 7.3125 234 1.2457 0.5824 1.2457 1.1161
No log 7.375 236 1.3648 0.5914 1.3648 1.1682
No log 7.4375 238 1.1920 0.5890 1.1920 1.0918
No log 7.5 240 1.1083 0.5806 1.1083 1.0527
No log 7.5625 242 1.1347 0.56 1.1347 1.0652
No log 7.625 244 1.0707 0.6087 1.0707 1.0348
No log 7.6875 246 0.9602 0.6176 0.9602 0.9799
No log 7.75 248 0.8862 0.6809 0.8862 0.9414
No log 7.8125 250 0.8730 0.7027 0.8730 0.9343
No log 7.875 252 0.9684 0.6582 0.9684 0.9841
No log 7.9375 254 1.0899 0.6552 1.0899 1.0440
No log 8.0 256 1.2261 0.5922 1.2261 1.1073
No log 8.0625 258 1.2149 0.6087 1.2149 1.1022
No log 8.125 260 1.2175 0.6054 1.2175 1.1034
No log 8.1875 262 1.3732 0.6042 1.3732 1.1718
No log 8.25 264 1.3535 0.6283 1.3535 1.1634
No log 8.3125 266 1.5487 0.5684 1.5487 1.2445
No log 8.375 268 1.6853 0.5436 1.6853 1.2982
No log 8.4375 270 1.7105 0.5389 1.7105 1.3078
No log 8.5 272 1.3727 0.6044 1.3727 1.1716
No log 8.5625 274 1.1338 0.5697 1.1338 1.0648
No log 8.625 276 1.1147 0.5976 1.1147 1.0558
No log 8.6875 278 1.2746 0.6180 1.2746 1.1290
No log 8.75 280 1.4455 0.5514 1.4455 1.2023
No log 8.8125 282 1.4280 0.5895 1.4280 1.1950
No log 8.875 284 1.2703 0.6011 1.2703 1.1271
No log 8.9375 286 1.0751 0.6316 1.0751 1.0368
No log 9.0 288 1.0102 0.6667 1.0102 1.0051
No log 9.0625 290 1.0715 0.6316 1.0715 1.0351
No log 9.125 292 1.0947 0.6118 1.0947 1.0463
No log 9.1875 294 0.9577 0.6456 0.9577 0.9786
No log 9.25 296 0.9206 0.6759 0.9206 0.9595
No log 9.3125 298 0.8681 0.7083 0.8681 0.9317
No log 9.375 300 0.9160 0.6849 0.9160 0.9571
No log 9.4375 302 1.1056 0.6386 1.1056 1.0515
No log 9.5 304 1.4446 0.5667 1.4446 1.2019
No log 9.5625 306 1.6241 0.5579 1.6241 1.2744
No log 9.625 308 1.5094 0.5654 1.5094 1.2286
No log 9.6875 310 1.2048 0.625 1.2048 1.0976
No log 9.75 312 1.0138 0.6860 1.0138 1.0069
No log 9.8125 314 0.8024 0.6923 0.8024 0.8958
No log 9.875 316 0.7944 0.6923 0.7944 0.8913
No log 9.9375 318 0.7985 0.6923 0.7985 0.8936
No log 10.0 320 0.8656 0.6623 0.8656 0.9304
No log 10.0625 322 1.0407 0.6584 1.0407 1.0201
No log 10.125 324 1.1891 0.5965 1.1891 1.0904
No log 10.1875 326 1.1931 0.6154 1.1931 1.0923
No log 10.25 328 1.0873 0.6867 1.0873 1.0427
No log 10.3125 330 0.9811 0.675 0.9811 0.9905
No log 10.375 332 0.9173 0.6710 0.9173 0.9578
No log 10.4375 334 0.9822 0.6708 0.9822 0.9911
No log 10.5 336 1.2140 0.6145 1.2140 1.1018
No log 10.5625 338 1.3784 0.6228 1.3784 1.1740
No log 10.625 340 1.6127 0.5455 1.6127 1.2699
No log 10.6875 342 1.6349 0.5587 1.6349 1.2786
No log 10.75 344 1.4327 0.5763 1.4327 1.1969
No log 10.8125 346 1.4761 0.5475 1.4761 1.2150
No log 10.875 348 1.7514 0.4839 1.7514 1.3234
No log 10.9375 350 1.9977 0.4734 1.9977 1.4134
No log 11.0 352 2.2249 0.4528 2.2249 1.4916
No log 11.0625 354 2.1630 0.4528 2.1630 1.4707
No log 11.125 356 2.1369 0.4645 2.1369 1.4618
No log 11.1875 358 1.8334 0.4878 1.8334 1.3540
No log 11.25 360 1.4212 0.5778 1.4212 1.1921
No log 11.3125 362 1.1268 0.6329 1.1268 1.0615
No log 11.375 364 1.0538 0.6275 1.0538 1.0265
No log 11.4375 366 1.1352 0.6275 1.1352 1.0655
No log 11.5 368 1.3748 0.5488 1.3748 1.1725
No log 11.5625 370 1.8425 0.4762 1.8425 1.3574
No log 11.625 372 2.2207 0.4369 2.2207 1.4902
No log 11.6875 374 2.2809 0.4444 2.2809 1.5103
No log 11.75 376 2.1463 0.4400 2.1463 1.4650
No log 11.8125 378 1.7408 0.4114 1.7408 1.3194
No log 11.875 380 1.3710 0.5342 1.3710 1.1709
No log 11.9375 382 1.1464 0.6143 1.1464 1.0707
No log 12.0 384 1.0739 0.6331 1.0739 1.0363
No log 12.0625 386 1.1036 0.6383 1.1036 1.0505
No log 12.125 388 1.2407 0.5732 1.2407 1.1139
No log 12.1875 390 1.4095 0.5238 1.4095 1.1872
No log 12.25 392 1.5104 0.4734 1.5104 1.2290
No log 12.3125 394 1.3780 0.5063 1.3780 1.1739
No log 12.375 396 1.1204 0.6241 1.1204 1.0585
No log 12.4375 398 0.9742 0.6759 0.9742 0.9870
No log 12.5 400 0.8998 0.6939 0.8998 0.9486
No log 12.5625 402 0.8880 0.6887 0.8880 0.9423
No log 12.625 404 0.9015 0.6933 0.9015 0.9495
No log 12.6875 406 1.0008 0.6711 1.0008 1.0004
No log 12.75 408 1.1956 0.6471 1.1956 1.0935
No log 12.8125 410 1.3244 0.6298 1.3244 1.1508
No log 12.875 412 1.4319 0.6203 1.4319 1.1966
No log 12.9375 414 1.4036 0.6162 1.4036 1.1847
No log 13.0 416 1.2331 0.6592 1.2331 1.1105
No log 13.0625 418 1.0741 0.6375 1.0741 1.0364
No log 13.125 420 0.9340 0.6316 0.9340 0.9664
No log 13.1875 422 0.8996 0.6316 0.8996 0.9484
No log 13.25 424 0.9358 0.5867 0.9358 0.9673
No log 13.3125 426 1.0531 0.5405 1.0531 1.0262
No log 13.375 428 1.1919 0.6087 1.1919 1.0917
No log 13.4375 430 1.2217 0.6303 1.2217 1.1053
No log 13.5 432 1.1593 0.6587 1.1593 1.0767
No log 13.5625 434 1.1340 0.6590 1.1340 1.0649
No log 13.625 436 1.0049 0.6667 1.0049 1.0024
No log 13.6875 438 0.9315 0.6626 0.9315 0.9651
No log 13.75 440 0.9824 0.6667 0.9824 0.9911
No log 13.8125 442 0.9740 0.6667 0.9740 0.9869
No log 13.875 444 1.0357 0.6543 1.0357 1.0177
No log 13.9375 446 1.1952 0.6286 1.1952 1.0932
No log 14.0 448 1.4234 0.6257 1.4234 1.1931
No log 14.0625 450 1.5171 0.6277 1.5171 1.2317
No log 14.125 452 1.4762 0.6257 1.4762 1.2150
No log 14.1875 454 1.2904 0.6328 1.2904 1.1360
No log 14.25 456 1.2076 0.6404 1.2076 1.0989
No log 14.3125 458 1.0554 0.6864 1.0554 1.0273
No log 14.375 460 1.0405 0.6864 1.0405 1.0200
No log 14.4375 462 1.0273 0.6826 1.0273 1.0136
No log 14.5 464 1.0855 0.6471 1.0855 1.0419
No log 14.5625 466 1.1683 0.6199 1.1683 1.0809
No log 14.625 468 1.1835 0.6286 1.1835 1.0879
No log 14.6875 470 1.1183 0.6286 1.1183 1.0575
No log 14.75 472 1.0282 0.6310 1.0282 1.0140
No log 14.8125 474 0.9517 0.6545 0.9517 0.9756
No log 14.875 476 0.8788 0.6826 0.8788 0.9374
No log 14.9375 478 0.7674 0.7389 0.7674 0.8760
No log 15.0 480 0.7146 0.7564 0.7146 0.8454
No log 15.0625 482 0.7029 0.7692 0.7029 0.8384
No log 15.125 484 0.7051 0.7692 0.7051 0.8397
No log 15.1875 486 0.7088 0.7692 0.7088 0.8419
No log 15.25 488 0.7459 0.7468 0.7459 0.8636
No log 15.3125 490 0.8704 0.6988 0.8704 0.9330
No log 15.375 492 1.0939 0.6592 1.0939 1.0459
No log 15.4375 494 1.3429 0.6114 1.3429 1.1588
No log 15.5 496 1.4033 0.6122 1.4033 1.1846
No log 15.5625 498 1.3533 0.6256 1.3533 1.1633
0.463 15.625 500 1.1565 0.6702 1.1565 1.0754
0.463 15.6875 502 0.9035 0.7093 0.9035 0.9505
0.463 15.75 504 0.7642 0.7179 0.7642 0.8742
0.463 15.8125 506 0.7328 0.7333 0.7328 0.8561
0.463 15.875 508 0.7288 0.7619 0.7288 0.8537
0.463 15.9375 510 0.7492 0.7413 0.7492 0.8656
0.463 16.0 512 0.8078 0.6755 0.8078 0.8988
0.463 16.0625 514 0.8845 0.6133 0.8845 0.9405
0.463 16.125 516 0.9767 0.6405 0.9767 0.9883
0.463 16.1875 518 1.0463 0.6144 1.0463 1.0229
0.463 16.25 520 1.0063 0.6053 1.0063 1.0031
0.463 16.3125 522 0.9704 0.6494 0.9704 0.9851

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k7_task1_organization

Finetuned
(4019)
this model