ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k14_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8016
  • Qwk: 0.5875
  • Mse: 0.8016
  • Rmse: 0.8953

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0290 2 5.4249 -0.0279 5.4249 2.3291
No log 0.0580 4 3.1351 0.0809 3.1351 1.7706
No log 0.0870 6 2.2720 -0.0191 2.2720 1.5073
No log 0.1159 8 1.5480 0.1675 1.5480 1.2442
No log 0.1449 10 1.3468 0.1025 1.3468 1.1605
No log 0.1739 12 1.1711 0.2615 1.1711 1.0822
No log 0.2029 14 1.3415 0.2042 1.3415 1.1583
No log 0.2319 16 1.6974 0.0529 1.6974 1.3028
No log 0.2609 18 1.8524 0.0882 1.8524 1.3610
No log 0.2899 20 1.6701 0.0707 1.6701 1.2923
No log 0.3188 22 1.4300 0.1856 1.4300 1.1958
No log 0.3478 24 1.2810 0.2872 1.2810 1.1318
No log 0.3768 26 1.3519 0.1290 1.3519 1.1627
No log 0.4058 28 1.4513 0.0871 1.4513 1.2047
No log 0.4348 30 1.3979 0.1281 1.3979 1.1823
No log 0.4638 32 1.1809 0.2701 1.1809 1.0867
No log 0.4928 34 1.1350 0.2653 1.1350 1.0653
No log 0.5217 36 1.1800 0.2425 1.1800 1.0863
No log 0.5507 38 1.1656 0.2491 1.1656 1.0796
No log 0.5797 40 1.1906 0.2656 1.1906 1.0911
No log 0.6087 42 1.2327 0.2769 1.2327 1.1103
No log 0.6377 44 1.3246 0.1940 1.3246 1.1509
No log 0.6667 46 1.3957 0.1681 1.3957 1.1814
No log 0.6957 48 1.2756 0.2230 1.2756 1.1294
No log 0.7246 50 1.1143 0.3467 1.1143 1.0556
No log 0.7536 52 1.0760 0.3264 1.0760 1.0373
No log 0.7826 54 1.0461 0.3264 1.0461 1.0228
No log 0.8116 56 1.0347 0.3953 1.0347 1.0172
No log 0.8406 58 1.0958 0.4200 1.0958 1.0468
No log 0.8696 60 1.1746 0.3565 1.1746 1.0838
No log 0.8986 62 1.0790 0.3971 1.0790 1.0388
No log 0.9275 64 1.0017 0.3928 1.0017 1.0008
No log 0.9565 66 0.9557 0.3885 0.9557 0.9776
No log 0.9855 68 0.9681 0.3743 0.9681 0.9839
No log 1.0145 70 1.0274 0.3790 1.0274 1.0136
No log 1.0435 72 1.0307 0.4336 1.0307 1.0152
No log 1.0725 74 1.0191 0.3816 1.0191 1.0095
No log 1.1014 76 1.0714 0.3584 1.0714 1.0351
No log 1.1304 78 1.2168 0.3505 1.2168 1.1031
No log 1.1594 80 1.0816 0.3622 1.0816 1.0400
No log 1.1884 82 0.9658 0.4033 0.9658 0.9828
No log 1.2174 84 0.9632 0.4489 0.9632 0.9814
No log 1.2464 86 0.9654 0.4306 0.9654 0.9825
No log 1.2754 88 0.9762 0.4390 0.9762 0.9881
No log 1.3043 90 0.9662 0.4360 0.9662 0.9829
No log 1.3333 92 0.9437 0.3667 0.9437 0.9714
No log 1.3623 94 0.9229 0.4369 0.9229 0.9607
No log 1.3913 96 0.8554 0.4289 0.8554 0.9249
No log 1.4203 98 0.8261 0.5057 0.8261 0.9089
No log 1.4493 100 0.8048 0.5056 0.8048 0.8971
No log 1.4783 102 0.7931 0.5198 0.7931 0.8905
No log 1.5072 104 0.9835 0.4942 0.9835 0.9917
No log 1.5362 106 1.1200 0.4760 1.1200 1.0583
No log 1.5652 108 1.1114 0.4759 1.1114 1.0542
No log 1.5942 110 1.0276 0.5057 1.0276 1.0137
No log 1.6232 112 0.9423 0.5353 0.9423 0.9707
No log 1.6522 114 0.9125 0.5768 0.9125 0.9553
No log 1.6812 116 0.8779 0.5809 0.8779 0.9370
No log 1.7101 118 0.9129 0.5671 0.9129 0.9555
No log 1.7391 120 0.9114 0.5599 0.9114 0.9547
No log 1.7681 122 0.9924 0.5284 0.9924 0.9962
No log 1.7971 124 1.1557 0.4590 1.1557 1.0751
No log 1.8261 126 1.1494 0.4645 1.1494 1.0721
No log 1.8551 128 1.1180 0.4660 1.1180 1.0573
No log 1.8841 130 1.0580 0.4831 1.0580 1.0286
No log 1.9130 132 0.9504 0.4785 0.9504 0.9749
No log 1.9420 134 0.9277 0.4792 0.9277 0.9632
No log 1.9710 136 1.0780 0.4380 1.0780 1.0383
No log 2.0 138 1.1584 0.4413 1.1584 1.0763
No log 2.0290 140 1.1256 0.4631 1.1256 1.0609
No log 2.0580 142 0.8681 0.5818 0.8681 0.9317
No log 2.0870 144 0.8189 0.5707 0.8189 0.9049
No log 2.1159 146 0.9674 0.5852 0.9674 0.9836
No log 2.1449 148 1.2772 0.4427 1.2772 1.1301
No log 2.1739 150 1.3300 0.4132 1.3300 1.1533
No log 2.2029 152 1.2448 0.4044 1.2448 1.1157
No log 2.2319 154 1.1427 0.3660 1.1427 1.0690
No log 2.2609 156 1.0702 0.3864 1.0702 1.0345
No log 2.2899 158 1.0555 0.4667 1.0555 1.0274
No log 2.3188 160 1.1204 0.4792 1.1204 1.0585
No log 2.3478 162 1.1630 0.4507 1.1630 1.0784
No log 2.3768 164 1.0978 0.4747 1.0978 1.0478
No log 2.4058 166 1.0506 0.4944 1.0506 1.0250
No log 2.4348 168 1.0015 0.5307 1.0015 1.0007
No log 2.4638 170 0.9142 0.5435 0.9142 0.9562
No log 2.4928 172 0.8936 0.5004 0.8936 0.9453
No log 2.5217 174 0.9141 0.5213 0.9141 0.9561
No log 2.5507 176 0.9739 0.4415 0.9739 0.9869
No log 2.5797 178 0.9869 0.4340 0.9869 0.9934
No log 2.6087 180 1.0042 0.4060 1.0042 1.0021
No log 2.6377 182 0.9710 0.4411 0.9710 0.9854
No log 2.6667 184 0.9547 0.4504 0.9547 0.9771
No log 2.6957 186 1.0101 0.4915 1.0101 1.0050
No log 2.7246 188 1.0590 0.4769 1.0590 1.0291
No log 2.7536 190 1.1475 0.4347 1.1475 1.0712
No log 2.7826 192 1.2572 0.3955 1.2572 1.1213
No log 2.8116 194 1.2919 0.4138 1.2919 1.1366
No log 2.8406 196 1.2535 0.3955 1.2535 1.1196
No log 2.8696 198 1.0711 0.4266 1.0711 1.0349
No log 2.8986 200 0.9061 0.5438 0.9061 0.9519
No log 2.9275 202 0.8490 0.6017 0.8490 0.9214
No log 2.9565 204 0.8736 0.5670 0.8736 0.9347
No log 2.9855 206 0.9748 0.5141 0.9748 0.9873
No log 3.0145 208 1.1480 0.4405 1.1480 1.0714
No log 3.0435 210 1.1953 0.4709 1.1953 1.0933
No log 3.0725 212 1.0854 0.5346 1.0854 1.0418
No log 3.1014 214 0.8945 0.6003 0.8945 0.9458
No log 3.1304 216 0.8123 0.6443 0.8123 0.9013
No log 3.1594 218 0.8138 0.6340 0.8138 0.9021
No log 3.1884 220 0.9319 0.5702 0.9319 0.9653
No log 3.2174 222 1.0845 0.5371 1.0845 1.0414
No log 3.2464 224 1.1240 0.5321 1.1240 1.0602
No log 3.2754 226 1.0207 0.4986 1.0207 1.0103
No log 3.3043 228 0.8899 0.4791 0.8899 0.9434
No log 3.3333 230 0.8196 0.4312 0.8196 0.9053
No log 3.3623 232 0.7958 0.5050 0.7958 0.8921
No log 3.3913 234 0.7942 0.5468 0.7942 0.8912
No log 3.4203 236 0.8317 0.5892 0.8317 0.9120
No log 3.4493 238 0.9310 0.5254 0.9310 0.9649
No log 3.4783 240 1.0201 0.5545 1.0201 1.0100
No log 3.5072 242 1.0264 0.5522 1.0264 1.0131
No log 3.5362 244 0.9646 0.5355 0.9646 0.9821
No log 3.5652 246 0.9218 0.6084 0.9218 0.9601
No log 3.5942 248 0.9171 0.5822 0.9171 0.9577
No log 3.6232 250 0.9410 0.5882 0.9410 0.9701
No log 3.6522 252 0.9330 0.5469 0.9330 0.9659
No log 3.6812 254 0.9585 0.5469 0.9585 0.9790
No log 3.7101 256 0.9339 0.5469 0.9339 0.9664
No log 3.7391 258 0.8838 0.5616 0.8838 0.9401
No log 3.7681 260 0.9251 0.5757 0.9251 0.9618
No log 3.7971 262 1.0730 0.5057 1.0730 1.0358
No log 3.8261 264 1.1223 0.5313 1.1223 1.0594
No log 3.8551 266 1.0153 0.5532 1.0153 1.0076
No log 3.8841 268 0.8655 0.6432 0.8655 0.9303
No log 3.9130 270 0.8171 0.6569 0.8171 0.9039
No log 3.9420 272 0.9075 0.5394 0.9075 0.9526
No log 3.9710 274 1.0954 0.4971 1.0954 1.0466
No log 4.0 276 1.2483 0.4832 1.2483 1.1173
No log 4.0290 278 1.1687 0.5000 1.1687 1.0810
No log 4.0580 280 1.0488 0.5313 1.0488 1.0241
No log 4.0870 282 1.1070 0.5295 1.1070 1.0522
No log 4.1159 284 1.3583 0.4797 1.3583 1.1655
No log 4.1449 286 1.3228 0.4398 1.3228 1.1501
No log 4.1739 288 1.1875 0.5186 1.1875 1.0897
No log 4.2029 290 1.1151 0.5186 1.1151 1.0560
No log 4.2319 292 1.0293 0.5419 1.0293 1.0145
No log 4.2609 294 0.9692 0.5621 0.9692 0.9845
No log 4.2899 296 0.9835 0.5390 0.9835 0.9917
No log 4.3188 298 1.0589 0.5133 1.0589 1.0290
No log 4.3478 300 1.2254 0.4720 1.2254 1.1070
No log 4.3768 302 1.1957 0.4773 1.1957 1.0935
No log 4.4058 304 1.2253 0.4898 1.2253 1.1069
No log 4.4348 306 1.0765 0.5483 1.0765 1.0376
No log 4.4638 308 0.9521 0.5988 0.9521 0.9757
No log 4.4928 310 0.9595 0.5988 0.9595 0.9795
No log 4.5217 312 1.0163 0.5871 1.0163 1.0081
No log 4.5507 314 1.1472 0.5317 1.1472 1.0711
No log 4.5797 316 1.1125 0.5627 1.1125 1.0547
No log 4.6087 318 1.1216 0.5753 1.1216 1.0590
No log 4.6377 320 1.2763 0.4889 1.2763 1.1297
No log 4.6667 322 1.3517 0.4761 1.3517 1.1626
No log 4.6957 324 1.3670 0.4879 1.3670 1.1692
No log 4.7246 326 1.3050 0.4865 1.3050 1.1423
No log 4.7536 328 1.0537 0.5523 1.0537 1.0265
No log 4.7826 330 0.8372 0.6704 0.8372 0.9150
No log 4.8116 332 0.8406 0.6695 0.8406 0.9169
No log 4.8406 334 1.0055 0.5637 1.0055 1.0028
No log 4.8696 336 1.1487 0.4952 1.1487 1.0718
No log 4.8986 338 1.0279 0.5270 1.0279 1.0139
No log 4.9275 340 0.8200 0.6215 0.8200 0.9055
No log 4.9565 342 0.7004 0.6376 0.7004 0.8369
No log 4.9855 344 0.7073 0.6376 0.7073 0.8410
No log 5.0145 346 0.7250 0.6531 0.7250 0.8514
No log 5.0435 348 0.7986 0.6575 0.7986 0.8936
No log 5.0725 350 0.9846 0.5620 0.9846 0.9923
No log 5.1014 352 1.1202 0.5261 1.1202 1.0584
No log 5.1304 354 1.0689 0.5328 1.0689 1.0339
No log 5.1594 356 0.9836 0.5337 0.9836 0.9918
No log 5.1884 358 0.9374 0.5500 0.9374 0.9682
No log 5.2174 360 1.0652 0.5299 1.0652 1.0321
No log 5.2464 362 1.3271 0.5094 1.3271 1.1520
No log 5.2754 364 1.2962 0.5116 1.2962 1.1385
No log 5.3043 366 1.2109 0.5080 1.2109 1.1004
No log 5.3333 368 1.1972 0.4853 1.1972 1.0942
No log 5.3623 370 1.1849 0.4974 1.1849 1.0885
No log 5.3913 372 1.0438 0.5337 1.0438 1.0217
No log 5.4203 374 0.9595 0.5730 0.9595 0.9795
No log 5.4493 376 0.9298 0.5845 0.9298 0.9643
No log 5.4783 378 0.9147 0.6074 0.9147 0.9564
No log 5.5072 380 0.8567 0.6400 0.8567 0.9256
No log 5.5362 382 0.9047 0.6424 0.9047 0.9512
No log 5.5652 384 1.0229 0.6023 1.0229 1.0114
No log 5.5942 386 1.2405 0.5463 1.2405 1.1138
No log 5.6232 388 1.2327 0.5369 1.2327 1.1103
No log 5.6522 390 1.0657 0.5199 1.0657 1.0323
No log 5.6812 392 0.8967 0.6043 0.8967 0.9469
No log 5.7101 394 0.7312 0.6601 0.7312 0.8551
No log 5.7391 396 0.6896 0.6426 0.6896 0.8304
No log 5.7681 398 0.7041 0.6335 0.7041 0.8391
No log 5.7971 400 0.7111 0.6680 0.7111 0.8433
No log 5.8261 402 0.7569 0.6736 0.7569 0.8700
No log 5.8551 404 0.7772 0.6664 0.7772 0.8816
No log 5.8841 406 0.8769 0.6366 0.8769 0.9364
No log 5.9130 408 0.9013 0.6331 0.9013 0.9494
No log 5.9420 410 0.8938 0.6393 0.8938 0.9454
No log 5.9710 412 0.8377 0.6046 0.8377 0.9153
No log 6.0 414 0.8132 0.5351 0.8132 0.9018
No log 6.0290 416 0.8018 0.5530 0.8018 0.8955
No log 6.0580 418 0.8417 0.6301 0.8417 0.9174
No log 6.0870 420 1.0522 0.5347 1.0522 1.0258
No log 6.1159 422 1.1732 0.4940 1.1732 1.0832
No log 6.1449 424 1.1622 0.5572 1.1622 1.0780
No log 6.1739 426 1.1601 0.5571 1.1601 1.0771
No log 6.2029 428 1.1935 0.5025 1.1935 1.0925
No log 6.2319 430 1.1682 0.5413 1.1682 1.0808
No log 6.2609 432 1.0787 0.5431 1.0787 1.0386
No log 6.2899 434 1.1295 0.5247 1.1295 1.0628
No log 6.3188 436 1.2077 0.5397 1.2077 1.0989
No log 6.3478 438 1.3604 0.4617 1.3604 1.1664
No log 6.3768 440 1.3819 0.4575 1.3819 1.1756
No log 6.4058 442 1.2781 0.4762 1.2781 1.1305
No log 6.4348 444 1.0882 0.4578 1.0882 1.0432
No log 6.4638 446 1.0030 0.4982 1.0030 1.0015
No log 6.4928 448 0.9084 0.5704 0.9084 0.9531
No log 6.5217 450 0.8015 0.6529 0.8015 0.8952
No log 6.5507 452 0.7763 0.6959 0.7763 0.8811
No log 6.5797 454 0.8597 0.6658 0.8597 0.9272
No log 6.6087 456 1.0117 0.6437 1.0117 1.0058
No log 6.6377 458 1.0972 0.6064 1.0972 1.0475
No log 6.6667 460 1.1600 0.5770 1.1600 1.0770
No log 6.6957 462 0.9518 0.6302 0.9518 0.9756
No log 6.7246 464 0.7639 0.6335 0.7639 0.8740
No log 6.7536 466 0.6914 0.6496 0.6914 0.8315
No log 6.7826 468 0.7493 0.6797 0.7493 0.8656
No log 6.8116 470 0.9669 0.6071 0.9669 0.9833
No log 6.8406 472 0.9859 0.5754 0.9859 0.9929
No log 6.8696 474 0.8752 0.6398 0.8752 0.9355
No log 6.8986 476 0.8931 0.6081 0.8931 0.9451
No log 6.9275 478 0.9843 0.5681 0.9843 0.9921
No log 6.9565 480 1.0977 0.5479 1.0977 1.0477
No log 6.9855 482 1.0589 0.5501 1.0589 1.0290
No log 7.0145 484 0.8911 0.5619 0.8911 0.9440
No log 7.0435 486 0.7765 0.6071 0.7765 0.8812
No log 7.0725 488 0.7529 0.6028 0.7529 0.8677
No log 7.1014 490 0.8046 0.5943 0.8046 0.8970
No log 7.1304 492 0.9151 0.5458 0.9151 0.9566
No log 7.1594 494 0.9502 0.5515 0.9502 0.9748
No log 7.1884 496 0.9404 0.5606 0.9404 0.9697
No log 7.2174 498 0.8912 0.5939 0.8912 0.9440
0.4691 7.2464 500 0.9031 0.6208 0.9031 0.9503
0.4691 7.2754 502 0.8810 0.6208 0.8810 0.9386
0.4691 7.3043 504 0.8426 0.6343 0.8426 0.9179
0.4691 7.3333 506 0.8891 0.6111 0.8891 0.9429
0.4691 7.3623 508 0.9708 0.6101 0.9708 0.9853
0.4691 7.3913 510 0.9046 0.6067 0.9046 0.9511
0.4691 7.4203 512 0.8451 0.6334 0.8451 0.9193
0.4691 7.4493 514 0.8163 0.6471 0.8163 0.9035
0.4691 7.4783 516 0.7960 0.6529 0.7960 0.8922
0.4691 7.5072 518 0.7665 0.6617 0.7665 0.8755
0.4691 7.5362 520 0.8115 0.6408 0.8115 0.9009
0.4691 7.5652 522 0.9580 0.5484 0.9580 0.9788
0.4691 7.5942 524 1.0051 0.5382 1.0051 1.0025
0.4691 7.6232 526 0.9731 0.5382 0.9731 0.9865
0.4691 7.6522 528 0.8847 0.5679 0.8847 0.9406
0.4691 7.6812 530 0.8016 0.5875 0.8016 0.8953

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k14_task1_organization

Finetuned
(4023)
this model