ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k19_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7528
  • Qwk: -0.1033
  • Mse: 0.7528
  • Rmse: 0.8677

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0392 2 3.8262 0.0094 3.8262 1.9561
No log 0.0784 4 2.0448 0.0504 2.0448 1.4300
No log 0.1176 6 1.7647 0.0304 1.7647 1.3284
No log 0.1569 8 2.2653 0.0076 2.2653 1.5051
No log 0.1961 10 1.8323 -0.0028 1.8323 1.3536
No log 0.2353 12 0.8634 -0.0870 0.8634 0.9292
No log 0.2745 14 0.7652 -0.0695 0.7652 0.8748
No log 0.3137 16 0.7526 -0.0695 0.7526 0.8675
No log 0.3529 18 0.8818 -0.0545 0.8818 0.9390
No log 0.3922 20 1.0065 -0.0435 1.0065 1.0033
No log 0.4314 22 1.0274 -0.0133 1.0274 1.0136
No log 0.4706 24 1.0621 -0.0193 1.0621 1.0306
No log 0.5098 26 0.8754 0.0207 0.8754 0.9356
No log 0.5490 28 0.7486 -0.0160 0.7486 0.8652
No log 0.5882 30 0.7842 -0.0331 0.7842 0.8856
No log 0.6275 32 1.0284 -0.0133 1.0284 1.0141
No log 0.6667 34 1.9703 -0.0085 1.9703 1.4037
No log 0.7059 36 1.9889 -0.0053 1.9889 1.4103
No log 0.7451 38 1.2092 0.0065 1.2092 1.0997
No log 0.7843 40 0.7730 -0.1233 0.7730 0.8792
No log 0.8235 42 0.8086 -0.1241 0.8086 0.8992
No log 0.8627 44 1.0242 -0.0435 1.0242 1.0121
No log 0.9020 46 1.0774 -0.0468 1.0774 1.0380
No log 0.9412 48 1.0904 -0.0468 1.0904 1.0442
No log 0.9804 50 0.9007 0.0134 0.9007 0.9490
No log 1.0196 52 0.7378 0.0 0.7378 0.8589
No log 1.0588 54 0.7251 -0.0551 0.7251 0.8515
No log 1.0980 56 0.7936 0.1097 0.7936 0.8908
No log 1.1373 58 1.3210 -0.0193 1.3210 1.1493
No log 1.1765 60 1.6936 0.0065 1.6936 1.3014
No log 1.2157 62 1.5780 -0.0193 1.5780 1.2562
No log 1.2549 64 1.0878 -0.0686 1.0878 1.0430
No log 1.2941 66 0.8381 0.1239 0.8381 0.9155
No log 1.3333 68 0.7309 0.0496 0.7309 0.8549
No log 1.3725 70 0.7579 0.1627 0.7579 0.8706
No log 1.4118 72 0.9578 -0.1282 0.9578 0.9787
No log 1.4510 74 1.0268 -0.1283 1.0268 1.0133
No log 1.4902 76 0.8012 0.1387 0.8012 0.8951
No log 1.5294 78 0.7441 0.0970 0.7441 0.8626
No log 1.5686 80 0.8189 0.1701 0.8189 0.9049
No log 1.6078 82 0.8377 0.1758 0.8377 0.9153
No log 1.6471 84 0.7729 0.0269 0.7729 0.8791
No log 1.6863 86 0.8590 0.1003 0.8590 0.9268
No log 1.7255 88 0.7640 -0.0145 0.7640 0.8741
No log 1.7647 90 0.7318 0.1027 0.7318 0.8555
No log 1.8039 92 0.7474 0.1501 0.7474 0.8645
No log 1.8431 94 0.8884 0.0257 0.8884 0.9426
No log 1.8824 96 0.8984 -0.0127 0.8984 0.9479
No log 1.9216 98 0.8115 0.1259 0.8115 0.9009
No log 1.9608 100 0.8449 0.1259 0.8449 0.9192
No log 2.0 102 0.9459 -0.0143 0.9459 0.9726
No log 2.0392 104 0.9856 0.0236 0.9856 0.9928
No log 2.0784 106 0.9755 0.1128 0.9755 0.9877
No log 2.1176 108 1.0471 0.0920 1.0471 1.0233
No log 2.1569 110 1.0316 0.1053 1.0316 1.0157
No log 2.1961 112 1.1321 0.0768 1.1321 1.0640
No log 2.2353 114 1.3865 -0.1226 1.3865 1.1775
No log 2.2745 116 1.1329 0.0 1.1329 1.0644
No log 2.3137 118 1.0967 0.0730 1.0967 1.0472
No log 2.3529 120 1.0843 -0.0231 1.0843 1.0413
No log 2.3922 122 1.2227 -0.1075 1.2227 1.1058
No log 2.4314 124 1.1372 -0.1439 1.1372 1.0664
No log 2.4706 126 0.9445 0.0392 0.9445 0.9718
No log 2.5098 128 0.8875 0.0771 0.8875 0.9421
No log 2.5490 130 1.0596 -0.0878 1.0596 1.0294
No log 2.5882 132 1.1874 -0.1276 1.1874 1.0897
No log 2.6275 134 0.9633 -0.0408 0.9633 0.9815
No log 2.6667 136 0.7796 -0.0032 0.7796 0.8830
No log 2.7059 138 0.7780 -0.1001 0.7780 0.8820
No log 2.7451 140 0.8060 -0.0145 0.8060 0.8978
No log 2.7843 142 1.0747 -0.0031 1.0747 1.0367
No log 2.8235 144 1.0334 -0.0031 1.0334 1.0166
No log 2.8627 146 0.8403 0.0226 0.8403 0.9167
No log 2.9020 148 0.7936 0.0893 0.7936 0.8908
No log 2.9412 150 0.8195 0.1199 0.8195 0.9053
No log 2.9804 152 0.8176 0.0680 0.8176 0.9042
No log 3.0196 154 0.7897 0.1199 0.7897 0.8887
No log 3.0588 156 0.8775 -0.0373 0.8775 0.9367
No log 3.0980 158 0.8251 0.0639 0.8251 0.9083
No log 3.1373 160 0.7424 0.0918 0.7424 0.8616
No log 3.1765 162 0.7598 0.1254 0.7598 0.8717
No log 3.2157 164 0.8395 0.1387 0.8395 0.9162
No log 3.2549 166 0.8078 0.0611 0.8078 0.8988
No log 3.2941 168 0.7995 0.0192 0.7995 0.8941
No log 3.3333 170 0.8608 0.1005 0.8608 0.9278
No log 3.3725 172 0.8836 0.0045 0.8836 0.9400
No log 3.4118 174 1.0633 0.0722 1.0633 1.0312
No log 3.4510 176 0.9783 -0.1606 0.9783 0.9891
No log 3.4902 178 0.9349 0.0966 0.9349 0.9669
No log 3.5294 180 0.9119 0.0964 0.9119 0.9549
No log 3.5686 182 1.0238 -0.1060 1.0238 1.0119
No log 3.6078 184 1.6216 -0.0400 1.6216 1.2734
No log 3.6471 186 1.5996 -0.0957 1.5996 1.2648
No log 3.6863 188 1.1263 -0.0194 1.1263 1.0613
No log 3.7255 190 0.8453 -0.0054 0.8453 0.9194
No log 3.7647 192 0.8203 0.0460 0.8203 0.9057
No log 3.8039 194 0.8546 -0.0303 0.8546 0.9245
No log 3.8431 196 1.0922 0.0486 1.0922 1.0451
No log 3.8824 198 1.1517 0.0778 1.1517 1.0732
No log 3.9216 200 0.8361 0.1742 0.8361 0.9144
No log 3.9608 202 0.7582 0.1637 0.7582 0.8708
No log 4.0 204 0.8419 0.2183 0.8419 0.9175
No log 4.0392 206 0.8173 0.1896 0.8173 0.9040
No log 4.0784 208 0.7327 0.1942 0.7327 0.8560
No log 4.1176 210 0.9036 0.2345 0.9036 0.9506
No log 4.1569 212 0.9571 0.1773 0.9571 0.9783
No log 4.1961 214 0.7813 0.1716 0.7813 0.8839
No log 4.2353 216 0.6735 0.1254 0.6735 0.8207
No log 4.2745 218 0.6960 0.0513 0.6960 0.8343
No log 4.3137 220 0.6762 0.1927 0.6762 0.8223
No log 4.3529 222 0.7360 0.1553 0.7360 0.8579
No log 4.3922 224 0.9112 0.0377 0.9112 0.9546
No log 4.4314 226 0.8738 0.0041 0.8738 0.9348
No log 4.4706 228 0.7896 0.1599 0.7896 0.8886
No log 4.5098 230 0.8274 0.1296 0.8274 0.9096
No log 4.5490 232 0.8262 0.1043 0.8262 0.9089
No log 4.5882 234 0.7381 0.1986 0.7381 0.8592
No log 4.6275 236 1.0077 0.1269 1.0077 1.0039
No log 4.6667 238 1.0109 0.0946 1.0109 1.0054
No log 4.7059 240 0.7264 0.2349 0.7264 0.8523
No log 4.7451 242 0.7600 0.0726 0.7600 0.8718
No log 4.7843 244 0.8647 -0.0120 0.8647 0.9299
No log 4.8235 246 0.7843 0.0345 0.7843 0.8856
No log 4.8627 248 0.6907 0.1807 0.6907 0.8311
No log 4.9020 250 0.9441 0.0651 0.9441 0.9716
No log 4.9412 252 1.1144 0.0634 1.1144 1.0556
No log 4.9804 254 0.9860 0.0487 0.9860 0.9930
No log 5.0196 256 0.8900 0.1135 0.8900 0.9434
No log 5.0588 258 0.8905 0.0753 0.8905 0.9436
No log 5.0980 260 0.9261 0.1066 0.9261 0.9623
No log 5.1373 262 1.0396 0.1110 1.0396 1.0196
No log 5.1765 264 0.9249 0.1065 0.9249 0.9617
No log 5.2157 266 0.8124 -0.0054 0.8124 0.9013
No log 5.2549 268 0.8027 0.0791 0.8027 0.8959
No log 5.2941 270 0.9841 0.0293 0.9841 0.9920
No log 5.3333 272 1.0535 -0.0211 1.0535 1.0264
No log 5.3725 274 0.9034 0.0871 0.9034 0.9505
No log 5.4118 276 0.7852 0.0821 0.7852 0.8861
No log 5.4510 278 0.7807 0.0874 0.7807 0.8836
No log 5.4902 280 0.7882 0.0884 0.7882 0.8878
No log 5.5294 282 0.8445 0.1485 0.8445 0.9190
No log 5.5686 284 0.8651 0.0140 0.8651 0.9301
No log 5.6078 286 0.8906 0.0562 0.8906 0.9437
No log 5.6471 288 0.9210 0.0821 0.9210 0.9597
No log 5.6863 290 1.0182 0.1297 1.0182 1.0091
No log 5.7255 292 0.9115 0.1379 0.9115 0.9547
No log 5.7647 294 0.8332 0.1899 0.8332 0.9128
No log 5.8039 296 0.7687 0.1141 0.7687 0.8768
No log 5.8431 298 0.7816 0.1440 0.7816 0.8841
No log 5.8824 300 0.9800 0.0182 0.9800 0.9900
No log 5.9216 302 1.0105 0.0873 1.0105 1.0052
No log 5.9608 304 0.7882 0.2424 0.7882 0.8878
No log 6.0 306 0.7375 -0.0407 0.7375 0.8588
No log 6.0392 308 0.7635 -0.0774 0.7635 0.8738
No log 6.0784 310 0.7438 0.0488 0.7438 0.8625
No log 6.1176 312 0.7710 0.1553 0.7710 0.8781
No log 6.1569 314 0.8580 0.0377 0.8580 0.9263
No log 6.1961 316 0.7890 0.0476 0.7890 0.8883
No log 6.2353 318 0.7545 0.2034 0.7545 0.8686
No log 6.2745 320 0.7431 0.2105 0.7431 0.8620
No log 6.3137 322 0.8104 0.0017 0.8104 0.9002
No log 6.3529 324 0.8176 0.0017 0.8176 0.9042
No log 6.3922 326 0.7890 0.1965 0.7890 0.8882
No log 6.4314 328 0.7685 0.0394 0.7685 0.8767
No log 6.4706 330 0.8216 0.1879 0.8216 0.9064
No log 6.5098 332 1.0344 0.0723 1.0344 1.0171
No log 6.5490 334 1.0376 0.0428 1.0376 1.0186
No log 6.5882 336 0.9211 0.1484 0.9211 0.9598
No log 6.6275 338 0.8672 0.2092 0.8672 0.9312
No log 6.6667 340 1.0082 0.1113 1.0082 1.0041
No log 6.7059 342 1.2156 -0.0133 1.2156 1.1026
No log 6.7451 344 1.0210 0.1149 1.0210 1.0105
No log 6.7843 346 0.8217 0.0749 0.8217 0.9065
No log 6.8235 348 0.8188 0.0798 0.8188 0.9049
No log 6.8627 350 0.9811 0.0576 0.9811 0.9905
No log 6.9020 352 1.3354 -0.0422 1.3354 1.1556
No log 6.9412 354 1.2028 0.0212 1.2028 1.0967
No log 6.9804 356 0.8327 0.1095 0.8327 0.9125
No log 7.0196 358 0.7922 -0.0370 0.7922 0.8901
No log 7.0588 360 0.7942 -0.0407 0.7942 0.8912
No log 7.0980 362 0.7849 0.0488 0.7849 0.8859
No log 7.1373 364 0.8237 0.0783 0.8237 0.9076
No log 7.1765 366 0.9085 0.1188 0.9085 0.9531
No log 7.2157 368 0.8345 0.0 0.8345 0.9135
No log 7.2549 370 0.8249 0.0145 0.8249 0.9082
No log 7.2941 372 0.8168 -0.0320 0.8168 0.9037
No log 7.3333 374 0.8223 0.1354 0.8223 0.9068
No log 7.3725 376 0.8296 0.0574 0.8296 0.9108
No log 7.4118 378 0.8130 0.0611 0.8130 0.9017
No log 7.4510 380 0.8966 0.0041 0.8966 0.9469
No log 7.4902 382 0.9006 0.0041 0.9006 0.9490
No log 7.5294 384 0.8508 0.0123 0.8508 0.9224
No log 7.5686 386 0.8372 0.0481 0.8372 0.9150
No log 7.6078 388 0.8432 0.0081 0.8432 0.9182
No log 7.6471 390 0.8701 0.1143 0.8701 0.9328
No log 7.6863 392 1.0193 -0.0073 1.0193 1.0096
No log 7.7255 394 0.9759 -0.0409 0.9759 0.9879
No log 7.7647 396 0.8917 0.0600 0.8917 0.9443
No log 7.8039 398 0.8696 0.1144 0.8696 0.9325
No log 7.8431 400 0.8988 0.1379 0.8988 0.9481
No log 7.8824 402 0.9929 -0.0094 0.9929 0.9964
No log 7.9216 404 1.0579 -0.0896 1.0579 1.0286
No log 7.9608 406 0.9568 -0.0073 0.9568 0.9782
No log 8.0 408 0.8821 -0.0743 0.8821 0.9392
No log 8.0392 410 0.8475 0.0247 0.8475 0.9206
No log 8.0784 412 0.8932 0.0214 0.8932 0.9451
No log 8.1176 414 1.0516 -0.0090 1.0516 1.0255
No log 8.1569 416 1.1971 -0.0870 1.1971 1.0941
No log 8.1961 418 1.1249 -0.0516 1.1249 1.0606
No log 8.2353 420 0.8932 0.0123 0.8932 0.9451
No log 8.2745 422 0.8368 -0.0179 0.8368 0.9148
No log 8.3137 424 0.8622 -0.0295 0.8622 0.9286
No log 8.3529 426 1.1080 -0.0597 1.1080 1.0526
No log 8.3922 428 1.6410 -0.0422 1.6410 1.2810
No log 8.4314 430 1.6845 -0.0187 1.6845 1.2979
No log 8.4706 432 1.1465 -0.0252 1.1465 1.0707
No log 8.5098 434 0.8287 -0.0179 0.8287 0.9104
No log 8.5490 436 0.8100 -0.0550 0.8100 0.9000
No log 8.5882 438 0.8035 -0.0660 0.8035 0.8964
No log 8.6275 440 0.8745 0.1775 0.8745 0.9352
No log 8.6667 442 1.0209 -0.0228 1.0209 1.0104
No log 8.7059 444 1.0605 -0.0228 1.0605 1.0298
No log 8.7451 446 0.9421 -0.0079 0.9421 0.9706
No log 8.7843 448 0.8352 -0.1333 0.8352 0.9139
No log 8.8235 450 0.8366 -0.0572 0.8366 0.9147
No log 8.8627 452 0.8462 -0.0806 0.8462 0.9199
No log 8.9020 454 0.9503 0.0362 0.9503 0.9748
No log 8.9412 456 0.9334 0.0793 0.9334 0.9661
No log 8.9804 458 0.9085 0.0793 0.9085 0.9531
No log 9.0196 460 0.8742 0.0793 0.8742 0.9350
No log 9.0588 462 0.8407 0.0549 0.8407 0.9169
No log 9.0980 464 0.8397 0.0628 0.8397 0.9164
No log 9.1373 466 0.8861 0.0831 0.8861 0.9413
No log 9.1765 468 0.9042 0.0504 0.9042 0.9509
No log 9.2157 470 0.8410 0.0700 0.8410 0.9171
No log 9.2549 472 0.8331 -0.1163 0.8331 0.9128
No log 9.2941 474 0.8412 0.0023 0.8412 0.9171
No log 9.3333 476 0.8587 0.1095 0.8587 0.9267
No log 9.3725 478 0.8471 0.1541 0.8471 0.9204
No log 9.4118 480 0.7996 0.1095 0.7996 0.8942
No log 9.4510 482 0.8007 0.0700 0.8007 0.8948
No log 9.4902 484 0.8532 0.1049 0.8532 0.9237
No log 9.5294 486 0.9868 0.1103 0.9868 0.9934
No log 9.5686 488 1.1467 0.0829 1.1467 1.0708
No log 9.6078 490 1.0672 0.0623 1.0672 1.0331
No log 9.6471 492 0.9285 0.0903 0.9285 0.9636
No log 9.6863 494 0.8952 0.0087 0.8952 0.9461
No log 9.7255 496 0.8529 0.0047 0.8529 0.9235
No log 9.7647 498 0.8076 0.0709 0.8076 0.8986
0.3326 9.8039 500 0.8422 0.1727 0.8422 0.9177
0.3326 9.8431 502 1.0515 0.1703 1.0515 1.0255
0.3326 9.8824 504 1.2547 0.0441 1.2547 1.1201
0.3326 9.9216 506 1.1386 0.0006 1.1386 1.0670
0.3326 9.9608 508 0.8144 0.1449 0.8144 0.9025
0.3326 10.0 510 0.7528 -0.1033 0.7528 0.8677

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k19_task3_organization

Finetuned
(4023)
this model