ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k4_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9833
  • Qwk: 0.6711
  • Mse: 0.9833
  • Rmse: 0.9916

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1053 2 7.0221 0.0179 7.0221 2.6499
No log 0.2105 4 4.5949 0.0599 4.5949 2.1436
No log 0.3158 6 3.5183 -0.0109 3.5183 1.8757
No log 0.4211 8 3.1243 0.0370 3.1243 1.7676
No log 0.5263 10 2.3873 0.1045 2.3873 1.5451
No log 0.6316 12 1.8805 0.1062 1.8805 1.3713
No log 0.7368 14 1.8449 0.1636 1.8449 1.3583
No log 0.8421 16 1.9785 0.1754 1.9785 1.4066
No log 0.9474 18 2.0095 0.1709 2.0095 1.4176
No log 1.0526 20 1.8287 0.2564 1.8287 1.3523
No log 1.1579 22 1.9336 0.1760 1.9336 1.3905
No log 1.2632 24 2.0701 0.1926 2.0701 1.4388
No log 1.3684 26 2.2158 0.1143 2.2158 1.4886
No log 1.4737 28 2.1055 0.1898 2.1055 1.4510
No log 1.5789 30 1.8739 0.1890 1.8739 1.3689
No log 1.6842 32 2.1622 0.1538 2.1622 1.4705
No log 1.7895 34 2.4460 0.1324 2.4460 1.5640
No log 1.8947 36 2.0974 0.2206 2.0974 1.4482
No log 2.0 38 1.7791 0.2857 1.7791 1.3338
No log 2.1053 40 1.6132 0.3969 1.6132 1.2701
No log 2.2105 42 1.4399 0.3193 1.4399 1.1999
No log 2.3158 44 1.3494 0.3932 1.3494 1.1616
No log 2.4211 46 1.3752 0.4034 1.3752 1.1727
No log 2.5263 48 1.2945 0.4167 1.2945 1.1378
No log 2.6316 50 1.1513 0.4878 1.1513 1.0730
No log 2.7368 52 1.3037 0.4545 1.3037 1.1418
No log 2.8421 54 1.3577 0.4662 1.3577 1.1652
No log 2.9474 56 1.2098 0.5909 1.2098 1.0999
No log 3.0526 58 1.1859 0.5758 1.1859 1.0890
No log 3.1579 60 1.1055 0.5581 1.1055 1.0514
No log 3.2632 62 1.0794 0.5354 1.0794 1.0390
No log 3.3684 64 1.0747 0.5581 1.0747 1.0367
No log 3.4737 66 1.1846 0.5 1.1846 1.0884
No log 3.5789 68 1.1838 0.5469 1.1838 1.0880
No log 3.6842 70 1.1717 0.5426 1.1717 1.0824
No log 3.7895 72 1.2330 0.5312 1.2330 1.1104
No log 3.8947 74 1.3453 0.4640 1.3453 1.1599
No log 4.0 76 1.2117 0.5606 1.2117 1.1008
No log 4.1053 78 1.1241 0.5152 1.1241 1.0602
No log 4.2105 80 1.1164 0.5802 1.1164 1.0566
No log 4.3158 82 1.1045 0.5758 1.1045 1.0509
No log 4.4211 84 1.2018 0.5231 1.2018 1.0963
No log 4.5263 86 1.2661 0.4762 1.2661 1.1252
No log 4.6316 88 1.2317 0.5231 1.2317 1.1098
No log 4.7368 90 1.2629 0.5191 1.2629 1.1238
No log 4.8421 92 1.2375 0.5224 1.2375 1.1124
No log 4.9474 94 1.3787 0.4755 1.3787 1.1742
No log 5.0526 96 1.6353 0.3889 1.6353 1.2788
No log 5.1579 98 1.7929 0.3699 1.7929 1.3390
No log 5.2632 100 1.4223 0.4892 1.4223 1.1926
No log 5.3684 102 1.0622 0.6176 1.0622 1.0306
No log 5.4737 104 1.0069 0.6176 1.0069 1.0035
No log 5.5789 106 0.9760 0.6176 0.9760 0.9879
No log 5.6842 108 0.9925 0.6569 0.9925 0.9963
No log 5.7895 110 1.0761 0.6377 1.0761 1.0374
No log 5.8947 112 1.1230 0.6377 1.1230 1.0597
No log 6.0 114 1.0767 0.6522 1.0767 1.0376
No log 6.1053 116 1.0486 0.6619 1.0486 1.0240
No log 6.2105 118 1.0993 0.6143 1.0993 1.0485
No log 6.3158 120 1.1579 0.5816 1.1579 1.0761
No log 6.4211 122 1.2394 0.4925 1.2394 1.1133
No log 6.5263 124 1.2215 0.5588 1.2215 1.1052
No log 6.6316 126 1.1698 0.5303 1.1698 1.0816
No log 6.7368 128 1.0858 0.6087 1.0858 1.0420
No log 6.8421 130 1.0502 0.6176 1.0502 1.0248
No log 6.9474 132 1.0673 0.6131 1.0673 1.0331
No log 7.0526 134 1.0845 0.6176 1.0845 1.0414
No log 7.1579 136 1.1077 0.5588 1.1077 1.0525
No log 7.2632 138 1.1022 0.5882 1.1022 1.0499
No log 7.3684 140 1.0812 0.5882 1.0812 1.0398
No log 7.4737 142 1.0720 0.6176 1.0720 1.0354
No log 7.5789 144 1.1621 0.6029 1.1621 1.0780
No log 7.6842 146 1.4076 0.4818 1.4076 1.1864
No log 7.7895 148 1.4850 0.4559 1.4850 1.2186
No log 7.8947 150 1.4102 0.4493 1.4102 1.1875
No log 8.0 152 1.1125 0.5652 1.1125 1.0548
No log 8.1053 154 1.0316 0.5942 1.0316 1.0157
No log 8.2105 156 1.0943 0.5839 1.0943 1.0461
No log 8.3158 158 1.1497 0.5672 1.1497 1.0722
No log 8.4211 160 1.0931 0.5839 1.0931 1.0455
No log 8.5263 162 1.1026 0.5672 1.1026 1.0501
No log 8.6316 164 1.1066 0.5882 1.1066 1.0519
No log 8.7368 166 1.0888 0.5882 1.0888 1.0435
No log 8.8421 168 1.0366 0.6087 1.0366 1.0181
No log 8.9474 170 0.9769 0.6571 0.9769 0.9884
No log 9.0526 172 0.9570 0.6571 0.9570 0.9782
No log 9.1579 174 0.9802 0.6429 0.9802 0.9901
No log 9.2632 176 1.0077 0.6429 1.0077 1.0039
No log 9.3684 178 1.0362 0.6571 1.0362 1.0180
No log 9.4737 180 0.9415 0.7273 0.9415 0.9703
No log 9.5789 182 0.8701 0.7273 0.8701 0.9328
No log 9.6842 184 0.8939 0.6950 0.8939 0.9454
No log 9.7895 186 0.9759 0.6429 0.9759 0.9879
No log 9.8947 188 1.1165 0.5957 1.1165 1.0566
No log 10.0 190 1.1982 0.5833 1.1982 1.0946
No log 10.1053 192 1.2082 0.5072 1.2082 1.0992
No log 10.2105 194 1.1477 0.5401 1.1477 1.0713
No log 10.3158 196 1.0936 0.6074 1.0936 1.0458
No log 10.4211 198 1.0993 0.6222 1.0993 1.0485
No log 10.5263 200 1.1281 0.5839 1.1281 1.0621
No log 10.6316 202 1.1596 0.5839 1.1596 1.0769
No log 10.7368 204 1.1928 0.5564 1.1928 1.0922
No log 10.8421 206 1.3053 0.5175 1.3053 1.1425
No log 10.9474 208 1.4898 0.5205 1.4898 1.2206
No log 11.0526 210 1.4717 0.4507 1.4717 1.2131
No log 11.1579 212 1.3588 0.4706 1.3588 1.1657
No log 11.2632 214 1.2904 0.5401 1.2904 1.1360
No log 11.3684 216 1.2295 0.5547 1.2295 1.1088
No log 11.4737 218 1.1935 0.5373 1.1935 1.0925
No log 11.5789 220 1.1461 0.5839 1.1461 1.0706
No log 11.6842 222 1.1302 0.5588 1.1302 1.0631
No log 11.7895 224 1.1164 0.5839 1.1164 1.0566
No log 11.8947 226 1.1432 0.5612 1.1432 1.0692
No log 12.0 228 1.1340 0.5362 1.1340 1.0649
No log 12.1053 230 1.0827 0.5882 1.0827 1.0405
No log 12.2105 232 1.0465 0.5821 1.0465 1.0230
No log 12.3158 234 1.0508 0.6074 1.0508 1.0251
No log 12.4211 236 1.0421 0.6087 1.0421 1.0208
No log 12.5263 238 1.0584 0.6043 1.0584 1.0288
No log 12.6316 240 1.0525 0.6043 1.0525 1.0259
No log 12.7368 242 1.0033 0.6423 1.0033 1.0016
No log 12.8421 244 0.9837 0.6901 0.9837 0.9918
No log 12.9474 246 0.9496 0.7143 0.9496 0.9745
No log 13.0526 248 0.9507 0.6815 0.9507 0.9751
No log 13.1579 250 0.9449 0.7007 0.9449 0.9720
No log 13.2632 252 0.9578 0.6667 0.9578 0.9787
No log 13.3684 254 0.9634 0.6715 0.9634 0.9815
No log 13.4737 256 0.9680 0.6715 0.9680 0.9839
No log 13.5789 258 0.9455 0.6812 0.9455 0.9724
No log 13.6842 260 0.9543 0.6957 0.9543 0.9769
No log 13.7895 262 0.9634 0.6815 0.9634 0.9815
No log 13.8947 264 0.9518 0.6815 0.9518 0.9756
No log 14.0 266 0.9506 0.6815 0.9506 0.9750
No log 14.1053 268 0.9623 0.6861 0.9623 0.9810
No log 14.2105 270 0.9499 0.6765 0.9499 0.9747
No log 14.3158 272 0.9456 0.6667 0.9456 0.9724
No log 14.4211 274 0.9634 0.7059 0.9634 0.9815
No log 14.5263 276 0.9526 0.7101 0.9526 0.9760
No log 14.6316 278 0.9235 0.6618 0.9235 0.9610
No log 14.7368 280 0.9053 0.7143 0.9053 0.9515
No log 14.8421 282 0.8875 0.7143 0.8875 0.9421
No log 14.9474 284 0.8796 0.7092 0.8796 0.9378
No log 15.0526 286 0.8774 0.6857 0.8774 0.9367
No log 15.1579 288 0.9134 0.7092 0.9134 0.9557
No log 15.2632 290 0.9613 0.6571 0.9613 0.9805
No log 15.3684 292 1.0107 0.5714 1.0107 1.0053
No log 15.4737 294 0.9714 0.5839 0.9714 0.9856
No log 15.5789 296 0.9568 0.6131 0.9568 0.9781
No log 15.6842 298 0.9518 0.6619 0.9518 0.9756
No log 15.7895 300 0.9246 0.6857 0.9246 0.9615
No log 15.8947 302 0.9433 0.6957 0.9433 0.9712
No log 16.0 304 0.9843 0.6957 0.9843 0.9921
No log 16.1053 306 0.9971 0.6912 0.9971 0.9985
No log 16.2105 308 0.9755 0.6861 0.9755 0.9877
No log 16.3158 310 0.9608 0.6370 0.9608 0.9802
No log 16.4211 312 0.9747 0.6187 0.9747 0.9873
No log 16.5263 314 1.0119 0.5839 1.0119 1.0059
No log 16.6316 316 1.0195 0.5942 1.0195 1.0097
No log 16.7368 318 0.9626 0.6187 0.9626 0.9811
No log 16.8421 320 0.9152 0.6074 0.9152 0.9566
No log 16.9474 322 0.9332 0.6074 0.9332 0.9660
No log 17.0526 324 0.9892 0.5778 0.9892 0.9946
No log 17.1579 326 1.1101 0.5303 1.1101 1.0536
No log 17.2632 328 1.2287 0.5303 1.2287 1.1085
No log 17.3684 330 1.1359 0.5303 1.1359 1.0658
No log 17.4737 332 1.0019 0.5606 1.0019 1.0010
No log 17.5789 334 0.9090 0.6515 0.9090 0.9534
No log 17.6842 336 0.8925 0.6866 0.8925 0.9447
No log 17.7895 338 0.9072 0.6866 0.9072 0.9525
No log 17.8947 340 0.9447 0.6617 0.9447 0.9720
No log 18.0 342 1.0289 0.5692 1.0289 1.0143
No log 18.1053 344 1.0384 0.5802 1.0384 1.0190
No log 18.2105 346 0.9999 0.6061 0.9999 0.9999
No log 18.3158 348 0.9766 0.6316 0.9766 0.9882
No log 18.4211 350 0.9592 0.6316 0.9592 0.9794
No log 18.5263 352 0.9562 0.6222 0.9562 0.9779
No log 18.6316 354 0.9538 0.6423 0.9538 0.9766
No log 18.7368 356 0.9599 0.6277 0.9599 0.9798
No log 18.8421 358 0.9754 0.6176 0.9754 0.9876
No log 18.9474 360 0.9797 0.6176 0.9797 0.9898
No log 19.0526 362 1.0216 0.6377 1.0216 1.0107
No log 19.1579 364 1.0028 0.6176 1.0028 1.0014
No log 19.2632 366 0.9337 0.7007 0.9337 0.9663
No log 19.3684 368 0.9120 0.6567 0.9120 0.9550
No log 19.4737 370 0.9201 0.6567 0.9201 0.9592
No log 19.5789 372 0.9388 0.6466 0.9388 0.9689
No log 19.6842 374 0.9597 0.6815 0.9597 0.9796
No log 19.7895 376 0.9754 0.6260 0.9754 0.9876
No log 19.8947 378 1.0011 0.5802 1.0011 1.0005
No log 20.0 380 1.0621 0.5564 1.0621 1.0306
No log 20.1053 382 1.0944 0.5455 1.0944 1.0461
No log 20.2105 384 1.0620 0.5564 1.0620 1.0305
No log 20.3158 386 0.9949 0.5970 0.9949 0.9974
No log 20.4211 388 0.9841 0.6269 0.9841 0.9920
No log 20.5263 390 0.9961 0.6269 0.9961 0.9980
No log 20.6316 392 0.9937 0.6519 0.9937 0.9969
No log 20.7368 394 0.9887 0.6222 0.9887 0.9944
No log 20.8421 396 1.0331 0.5758 1.0331 1.0164
No log 20.9474 398 1.0698 0.5882 1.0698 1.0343
No log 21.0526 400 1.0632 0.6029 1.0632 1.0311
No log 21.1579 402 1.0174 0.6029 1.0174 1.0086
No log 21.2632 404 0.9694 0.5926 0.9694 0.9846
No log 21.3684 406 0.9770 0.5758 0.9770 0.9884
No log 21.4737 408 0.9908 0.6074 0.9908 0.9954
No log 21.5789 410 0.9982 0.5882 0.9982 0.9991
No log 21.6842 412 1.0022 0.5882 1.0022 1.0011
No log 21.7895 414 0.9563 0.6222 0.9563 0.9779
No log 21.8947 416 0.9155 0.6471 0.9155 0.9568
No log 22.0 418 0.9082 0.6232 0.9082 0.9530
No log 22.1053 420 0.8823 0.6522 0.8823 0.9393
No log 22.2105 422 0.8483 0.7143 0.8483 0.9210
No log 22.3158 424 0.8465 0.7338 0.8465 0.9201
No log 22.4211 426 0.8719 0.6906 0.8719 0.9338
No log 22.5263 428 0.9210 0.6619 0.9210 0.9597
No log 22.6316 430 0.9456 0.6131 0.9456 0.9724
No log 22.7368 432 0.9685 0.5882 0.9685 0.9841
No log 22.8421 434 0.9974 0.5564 0.9974 0.9987
No log 22.9474 436 0.9766 0.5797 0.9766 0.9882
No log 23.0526 438 0.9303 0.6857 0.9303 0.9645
No log 23.1579 440 0.9196 0.7092 0.9196 0.9590
No log 23.2632 442 0.9055 0.7092 0.9055 0.9516
No log 23.3684 444 0.8997 0.7 0.8997 0.9485
No log 23.4737 446 0.9252 0.6087 0.9252 0.9619
No log 23.5789 448 0.9389 0.6232 0.9389 0.9690
No log 23.6842 450 0.9348 0.6370 0.9348 0.9669
No log 23.7895 452 0.8876 0.6466 0.8876 0.9421
No log 23.8947 454 0.8490 0.6912 0.8490 0.9214
No log 24.0 456 0.8389 0.6963 0.8389 0.9159
No log 24.1053 458 0.8308 0.6912 0.8308 0.9115
No log 24.2105 460 0.8263 0.6963 0.8263 0.9090
No log 24.3158 462 0.8200 0.6912 0.8200 0.9055
No log 24.4211 464 0.8223 0.6963 0.8223 0.9068
No log 24.5263 466 0.8372 0.7413 0.8372 0.9150
No log 24.6316 468 0.8641 0.7413 0.8641 0.9296
No log 24.7368 470 0.9236 0.6714 0.9236 0.9611
No log 24.8421 472 0.9811 0.6187 0.9811 0.9905
No log 24.9474 474 0.9757 0.5942 0.9757 0.9878
No log 25.0526 476 0.9481 0.6525 0.9481 0.9737
No log 25.1579 478 0.8996 0.6806 0.8996 0.9485
No log 25.2632 480 0.8822 0.6897 0.8822 0.9392
No log 25.3684 482 0.8568 0.7222 0.8568 0.9256
No log 25.4737 484 0.8528 0.7222 0.8528 0.9235
No log 25.5789 486 0.8802 0.7133 0.8802 0.9382
No log 25.6842 488 0.9260 0.6571 0.9260 0.9623
No log 25.7895 490 0.9682 0.6331 0.9682 0.9840
No log 25.8947 492 0.9574 0.6331 0.9574 0.9785
No log 26.0 494 0.9596 0.6429 0.9596 0.9796
No log 26.1053 496 0.9384 0.6571 0.9384 0.9687
No log 26.2105 498 0.9078 0.6857 0.9078 0.9528
0.2866 26.3158 500 0.8964 0.7092 0.8964 0.9468
0.2866 26.4211 502 0.9122 0.7042 0.9122 0.9551
0.2866 26.5263 504 0.9693 0.6389 0.9693 0.9845
0.2866 26.6316 506 0.9839 0.6389 0.9839 0.9919
0.2866 26.7368 508 0.9552 0.7222 0.9552 0.9773
0.2866 26.8421 510 0.9907 0.6757 0.9907 0.9953
0.2866 26.9474 512 1.1398 0.6081 1.1398 1.0676
0.2866 27.0526 514 1.2639 0.6026 1.2639 1.1242
0.2866 27.1579 516 1.1542 0.5694 1.1542 1.0743
0.2866 27.2632 518 0.9833 0.6711 0.9833 0.9916

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k4_task1_organization

Finetuned
(4023)
this model