ArabicNewSplits5_FineTuningAraBERT_run3_AugV5_k9_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8242
  • Qwk: 0.6091
  • Mse: 0.8242
  • Rmse: 0.9079

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0345 2 5.1862 -0.0378 5.1862 2.2773
No log 0.0690 4 3.0577 0.0812 3.0577 1.7486
No log 0.1034 6 2.0758 0.0051 2.0758 1.4408
No log 0.1379 8 2.4292 -0.1028 2.4292 1.5586
No log 0.1724 10 2.1197 -0.0538 2.1197 1.4559
No log 0.2069 12 1.4985 0.0905 1.4985 1.2241
No log 0.2414 14 1.3001 0.2057 1.3001 1.1402
No log 0.2759 16 1.1769 0.2605 1.1769 1.0848
No log 0.3103 18 1.1800 0.3648 1.1800 1.0863
No log 0.3448 20 1.1733 0.3280 1.1733 1.0832
No log 0.3793 22 1.2307 0.3653 1.2307 1.1094
No log 0.4138 24 1.6613 0.2496 1.6613 1.2889
No log 0.4483 26 2.1120 0.2388 2.1120 1.4533
No log 0.4828 28 1.9788 0.2675 1.9788 1.4067
No log 0.5172 30 1.6182 0.3065 1.6182 1.2721
No log 0.5517 32 1.3496 0.2931 1.3496 1.1617
No log 0.5862 34 1.3186 0.3334 1.3186 1.1483
No log 0.6207 36 1.6070 0.2877 1.6070 1.2677
No log 0.6552 38 2.0197 0.2404 2.0197 1.4212
No log 0.6897 40 1.8706 0.2875 1.8706 1.3677
No log 0.7241 42 1.2809 0.3309 1.2809 1.1318
No log 0.7586 44 0.9582 0.4354 0.9582 0.9789
No log 0.7931 46 1.0200 0.3792 1.0200 1.0100
No log 0.8276 48 1.0736 0.3206 1.0736 1.0361
No log 0.8621 50 0.9901 0.3910 0.9901 0.9950
No log 0.8966 52 0.9590 0.4662 0.9590 0.9793
No log 0.9310 54 1.1032 0.4964 1.1032 1.0503
No log 0.9655 56 1.3656 0.3294 1.3656 1.1686
No log 1.0 58 1.8693 0.2750 1.8693 1.3672
No log 1.0345 60 2.0892 0.2645 2.0892 1.4454
No log 1.0690 62 2.0997 0.2595 2.0997 1.4490
No log 1.1034 64 1.9481 0.2444 1.9481 1.3957
No log 1.1379 66 1.5582 0.2982 1.5582 1.2483
No log 1.1724 68 1.2422 0.4276 1.2422 1.1145
No log 1.2069 70 1.0069 0.4864 1.0069 1.0034
No log 1.2414 72 0.9569 0.5010 0.9569 0.9782
No log 1.2759 74 0.9885 0.5349 0.9885 0.9942
No log 1.3103 76 0.9916 0.5301 0.9916 0.9958
No log 1.3448 78 1.0251 0.5215 1.0251 1.0125
No log 1.3793 80 1.0781 0.5541 1.0781 1.0383
No log 1.4138 82 1.1173 0.5613 1.1173 1.0570
No log 1.4483 84 1.2439 0.5503 1.2439 1.1153
No log 1.4828 86 1.4597 0.4959 1.4597 1.2082
No log 1.5172 88 1.3854 0.4884 1.3854 1.1770
No log 1.5517 90 1.0968 0.4999 1.0968 1.0473
No log 1.5862 92 1.0354 0.5004 1.0354 1.0175
No log 1.6207 94 0.9106 0.6163 0.9106 0.9542
No log 1.6552 96 0.8705 0.5996 0.8705 0.9330
No log 1.6897 98 0.8819 0.5991 0.8819 0.9391
No log 1.7241 100 0.8762 0.5658 0.8762 0.9361
No log 1.7586 102 0.8939 0.5758 0.8939 0.9455
No log 1.7931 104 0.9885 0.5706 0.9885 0.9942
No log 1.8276 106 1.0530 0.5321 1.0530 1.0262
No log 1.8621 108 1.1204 0.4646 1.1204 1.0585
No log 1.8966 110 1.1526 0.4234 1.1526 1.0736
No log 1.9310 112 1.0426 0.5480 1.0426 1.0211
No log 1.9655 114 0.9864 0.5704 0.9864 0.9932
No log 2.0 116 0.9486 0.5881 0.9486 0.9740
No log 2.0345 118 1.0775 0.5425 1.0775 1.0380
No log 2.0690 120 1.1878 0.4808 1.1878 1.0899
No log 2.1034 122 1.0194 0.5717 1.0194 1.0096
No log 2.1379 124 0.8319 0.5682 0.8319 0.9121
No log 2.1724 126 0.7993 0.6287 0.7993 0.8940
No log 2.2069 128 0.7969 0.6017 0.7969 0.8927
No log 2.2414 130 0.8220 0.6365 0.8220 0.9066
No log 2.2759 132 0.8184 0.6720 0.8184 0.9046
No log 2.3103 134 0.8653 0.6555 0.8653 0.9302
No log 2.3448 136 0.9262 0.5937 0.9262 0.9624
No log 2.3793 138 0.9743 0.5853 0.9743 0.9871
No log 2.4138 140 0.9900 0.5962 0.9900 0.9950
No log 2.4483 142 0.9976 0.6118 0.9976 0.9988
No log 2.4828 144 1.0002 0.6432 1.0002 1.0001
No log 2.5172 146 1.0366 0.6233 1.0366 1.0181
No log 2.5517 148 1.0498 0.5901 1.0498 1.0246
No log 2.5862 150 1.0991 0.5633 1.0991 1.0484
No log 2.6207 152 1.1327 0.5153 1.1327 1.0643
No log 2.6552 154 1.0076 0.5631 1.0076 1.0038
No log 2.6897 156 0.8848 0.5859 0.8848 0.9406
No log 2.7241 158 0.8704 0.6344 0.8704 0.9330
No log 2.7586 160 0.8758 0.6288 0.8758 0.9358
No log 2.7931 162 0.8692 0.6327 0.8692 0.9323
No log 2.8276 164 0.8413 0.6212 0.8413 0.9172
No log 2.8621 166 0.8576 0.5587 0.8576 0.9260
No log 2.8966 168 0.9691 0.5391 0.9691 0.9845
No log 2.9310 170 0.9275 0.5478 0.9275 0.9631
No log 2.9655 172 0.8227 0.5835 0.8227 0.9070
No log 3.0 174 0.7952 0.5973 0.7952 0.8918
No log 3.0345 176 0.8058 0.6346 0.8058 0.8977
No log 3.0690 178 0.8350 0.6656 0.8350 0.9138
No log 3.1034 180 0.8629 0.6502 0.8629 0.9289
No log 3.1379 182 0.8750 0.6254 0.8750 0.9354
No log 3.1724 184 0.8491 0.6206 0.8491 0.9215
No log 3.2069 186 0.8503 0.6138 0.8503 0.9221
No log 3.2414 188 0.8914 0.6342 0.8914 0.9441
No log 3.2759 190 0.9022 0.6320 0.9022 0.9498
No log 3.3103 192 0.8682 0.6172 0.8682 0.9318
No log 3.3448 194 0.8997 0.6183 0.8997 0.9485
No log 3.3793 196 0.9585 0.5990 0.9585 0.9790
No log 3.4138 198 0.9983 0.5997 0.9983 0.9992
No log 3.4483 200 0.9731 0.5824 0.9731 0.9864
No log 3.4828 202 0.8894 0.5772 0.8894 0.9431
No log 3.5172 204 0.8138 0.6292 0.8138 0.9021
No log 3.5517 206 0.7807 0.6866 0.7807 0.8836
No log 3.5862 208 0.7783 0.6681 0.7783 0.8822
No log 3.6207 210 0.8034 0.6582 0.8034 0.8963
No log 3.6552 212 0.8319 0.6367 0.8319 0.9121
No log 3.6897 214 0.8241 0.6469 0.8241 0.9078
No log 3.7241 216 0.7904 0.6456 0.7904 0.8890
No log 3.7586 218 0.7696 0.6693 0.7696 0.8773
No log 3.7931 220 0.7797 0.6668 0.7797 0.8830
No log 3.8276 222 0.8125 0.6337 0.8125 0.9014
No log 3.8621 224 0.8186 0.6308 0.8186 0.9048
No log 3.8966 226 0.8099 0.6239 0.8099 0.8999
No log 3.9310 228 0.8000 0.6230 0.8000 0.8944
No log 3.9655 230 0.8000 0.6542 0.8000 0.8944
No log 4.0 232 0.8077 0.6261 0.8077 0.8987
No log 4.0345 234 0.7874 0.6253 0.7874 0.8874
No log 4.0690 236 0.7857 0.6279 0.7857 0.8864
No log 4.1034 238 0.8154 0.6437 0.8154 0.9030
No log 4.1379 240 0.9026 0.6154 0.9026 0.9501
No log 4.1724 242 0.9798 0.5809 0.9798 0.9899
No log 4.2069 244 1.0225 0.5860 1.0225 1.0112
No log 4.2414 246 1.0003 0.5898 1.0003 1.0002
No log 4.2759 248 0.9371 0.5952 0.9371 0.9680
No log 4.3103 250 0.8468 0.6253 0.8468 0.9202
No log 4.3448 252 0.8051 0.6094 0.8051 0.8973
No log 4.3793 254 0.8051 0.5762 0.8051 0.8973
No log 4.4138 256 0.8181 0.5869 0.8181 0.9045
No log 4.4483 258 0.8486 0.5992 0.8486 0.9212
No log 4.4828 260 0.8767 0.5374 0.8767 0.9363
No log 4.5172 262 0.8821 0.5362 0.8821 0.9392
No log 4.5517 264 0.8767 0.5325 0.8767 0.9363
No log 4.5862 266 0.8595 0.5820 0.8595 0.9271
No log 4.6207 268 0.8531 0.5717 0.8531 0.9236
No log 4.6552 270 0.8257 0.5880 0.8257 0.9087
No log 4.6897 272 0.8056 0.6123 0.8056 0.8975
No log 4.7241 274 0.8271 0.6630 0.8271 0.9095
No log 4.7586 276 0.8815 0.6292 0.8815 0.9389
No log 4.7931 278 0.9323 0.6212 0.9323 0.9656
No log 4.8276 280 0.9271 0.5934 0.9271 0.9629
No log 4.8621 282 0.8849 0.5554 0.8849 0.9407
No log 4.8966 284 0.8742 0.5796 0.8742 0.9350
No log 4.9310 286 0.8645 0.6110 0.8645 0.9298
No log 4.9655 288 0.8370 0.6267 0.8370 0.9149
No log 5.0 290 0.8151 0.6442 0.8151 0.9028
No log 5.0345 292 0.8248 0.6245 0.8248 0.9082
No log 5.0690 294 0.8697 0.6321 0.8697 0.9326
No log 5.1034 296 0.9126 0.6272 0.9126 0.9553
No log 5.1379 298 0.9156 0.6071 0.9156 0.9569
No log 5.1724 300 0.8854 0.6149 0.8854 0.9409
No log 5.2069 302 0.8271 0.6239 0.8271 0.9094
No log 5.2414 304 0.7785 0.6433 0.7785 0.8823
No log 5.2759 306 0.7709 0.6435 0.7709 0.8780
No log 5.3103 308 0.7711 0.6586 0.7711 0.8781
No log 5.3448 310 0.7811 0.6382 0.7811 0.8838
No log 5.3793 312 0.7917 0.6209 0.7917 0.8898
No log 5.4138 314 0.7767 0.6441 0.7767 0.8813
No log 5.4483 316 0.7612 0.6443 0.7612 0.8725
No log 5.4828 318 0.7796 0.6292 0.7796 0.8829
No log 5.5172 320 0.7923 0.6477 0.7923 0.8901
No log 5.5517 322 0.8115 0.6160 0.8115 0.9008
No log 5.5862 324 0.8285 0.6434 0.8285 0.9102
No log 5.6207 326 0.8318 0.6461 0.8318 0.9120
No log 5.6552 328 0.8367 0.6555 0.8367 0.9147
No log 5.6897 330 0.8132 0.6239 0.8132 0.9018
No log 5.7241 332 0.7800 0.6636 0.7800 0.8832
No log 5.7586 334 0.7705 0.6686 0.7705 0.8778
No log 5.7931 336 0.7790 0.6428 0.7790 0.8826
No log 5.8276 338 0.7894 0.6567 0.7894 0.8885
No log 5.8621 340 0.8060 0.6427 0.8060 0.8978
No log 5.8966 342 0.8504 0.6190 0.8504 0.9221
No log 5.9310 344 0.8875 0.6063 0.8875 0.9421
No log 5.9655 346 0.9436 0.5996 0.9436 0.9714
No log 6.0 348 0.9954 0.5868 0.9954 0.9977
No log 6.0345 350 0.9874 0.5868 0.9874 0.9937
No log 6.0690 352 0.9185 0.6044 0.9185 0.9584
No log 6.1034 354 0.8326 0.6562 0.8326 0.9124
No log 6.1379 356 0.7821 0.6472 0.7821 0.8844
No log 6.1724 358 0.7695 0.6406 0.7695 0.8772
No log 6.2069 360 0.7661 0.6475 0.7661 0.8753
No log 6.2414 362 0.7745 0.6515 0.7745 0.8801
No log 6.2759 364 0.8067 0.6573 0.8067 0.8982
No log 6.3103 366 0.8578 0.6044 0.8578 0.9262
No log 6.3448 368 0.8921 0.6054 0.8921 0.9445
No log 6.3793 370 0.8529 0.6229 0.8529 0.9235
No log 6.4138 372 0.8159 0.6206 0.8159 0.9033
No log 6.4483 374 0.7780 0.6363 0.7780 0.8821
No log 6.4828 376 0.7502 0.6261 0.7502 0.8661
No log 6.5172 378 0.7478 0.6107 0.7478 0.8648
No log 6.5517 380 0.7484 0.6096 0.7484 0.8651
No log 6.5862 382 0.7496 0.5955 0.7496 0.8658
No log 6.6207 384 0.7624 0.5958 0.7624 0.8732
No log 6.6552 386 0.7876 0.5835 0.7876 0.8875
No log 6.6897 388 0.8064 0.5595 0.8064 0.8980
No log 6.7241 390 0.8212 0.5693 0.8212 0.9062
No log 6.7586 392 0.8182 0.5671 0.8182 0.9046
No log 6.7931 394 0.8005 0.5709 0.8005 0.8947
No log 6.8276 396 0.7845 0.5890 0.7845 0.8857
No log 6.8621 398 0.7878 0.6082 0.7878 0.8876
No log 6.8966 400 0.7900 0.6243 0.7900 0.8888
No log 6.9310 402 0.7952 0.6137 0.7952 0.8917
No log 6.9655 404 0.8012 0.6137 0.8012 0.8951
No log 7.0 406 0.8064 0.6054 0.8064 0.8980
No log 7.0345 408 0.8115 0.6012 0.8115 0.9008
No log 7.0690 410 0.8182 0.5762 0.8182 0.9045
No log 7.1034 412 0.8218 0.5748 0.8218 0.9065
No log 7.1379 414 0.8165 0.5770 0.8165 0.9036
No log 7.1724 416 0.8047 0.5787 0.8047 0.8971
No log 7.2069 418 0.8104 0.6099 0.8104 0.9002
No log 7.2414 420 0.8379 0.6212 0.8379 0.9154
No log 7.2759 422 0.8432 0.6318 0.8432 0.9183
No log 7.3103 424 0.8384 0.6221 0.8384 0.9157
No log 7.3448 426 0.8217 0.6220 0.8217 0.9065
No log 7.3793 428 0.7952 0.6406 0.7952 0.8917
No log 7.4138 430 0.7703 0.6372 0.7703 0.8777
No log 7.4483 432 0.7565 0.6420 0.7565 0.8698
No log 7.4828 434 0.7524 0.6527 0.7524 0.8674
No log 7.5172 436 0.7557 0.6594 0.7557 0.8693
No log 7.5517 438 0.7630 0.6386 0.7630 0.8735
No log 7.5862 440 0.7777 0.6197 0.7777 0.8818
No log 7.6207 442 0.7837 0.6169 0.7837 0.8852
No log 7.6552 444 0.7856 0.6347 0.7856 0.8863
No log 7.6897 446 0.7962 0.6302 0.7962 0.8923
No log 7.7241 448 0.8190 0.6263 0.8190 0.9050
No log 7.7586 450 0.8519 0.6337 0.8519 0.9230
No log 7.7931 452 0.8714 0.6541 0.8714 0.9335
No log 7.8276 454 0.8719 0.6541 0.8719 0.9337
No log 7.8621 456 0.8726 0.6541 0.8726 0.9341
No log 7.8966 458 0.8645 0.6337 0.8645 0.9298
No log 7.9310 460 0.8500 0.6085 0.8500 0.9220
No log 7.9655 462 0.8388 0.6134 0.8388 0.9159
No log 8.0 464 0.8372 0.6125 0.8372 0.9150
No log 8.0345 466 0.8354 0.6120 0.8354 0.9140
No log 8.0690 468 0.8356 0.6078 0.8356 0.9141
No log 8.1034 470 0.8429 0.5947 0.8429 0.9181
No log 8.1379 472 0.8561 0.5937 0.8561 0.9253
No log 8.1724 474 0.8688 0.5941 0.8688 0.9321
No log 8.2069 476 0.8688 0.5810 0.8688 0.9321
No log 8.2414 478 0.8654 0.5810 0.8654 0.9303
No log 8.2759 480 0.8597 0.6011 0.8597 0.9272
No log 8.3103 482 0.8424 0.6038 0.8424 0.9178
No log 8.3448 484 0.8268 0.6108 0.8268 0.9093
No log 8.3793 486 0.8026 0.6236 0.8026 0.8959
No log 8.4138 488 0.7883 0.6159 0.7883 0.8878
No log 8.4483 490 0.7820 0.6168 0.7820 0.8843
No log 8.4828 492 0.7802 0.6361 0.7802 0.8833
No log 8.5172 494 0.7802 0.6361 0.7802 0.8833
No log 8.5517 496 0.7827 0.6429 0.7827 0.8847
No log 8.5862 498 0.7904 0.6383 0.7904 0.8890
0.4074 8.6207 500 0.8078 0.6313 0.8078 0.8988
0.4074 8.6552 502 0.8245 0.6366 0.8245 0.9080
0.4074 8.6897 504 0.8368 0.6351 0.8368 0.9148
0.4074 8.7241 506 0.8520 0.6351 0.8520 0.9230
0.4074 8.7586 508 0.8606 0.6414 0.8606 0.9277
0.4074 8.7931 510 0.8532 0.6329 0.8532 0.9237
0.4074 8.8276 512 0.8381 0.6502 0.8381 0.9155
0.4074 8.8621 514 0.8213 0.6255 0.8213 0.9062
0.4074 8.8966 516 0.8023 0.6301 0.8023 0.8957
0.4074 8.9310 518 0.7874 0.6234 0.7874 0.8874
0.4074 8.9655 520 0.7773 0.6437 0.7773 0.8816
0.4074 9.0 522 0.7713 0.6386 0.7713 0.8782
0.4074 9.0345 524 0.7694 0.6386 0.7694 0.8772
0.4074 9.0690 526 0.7735 0.6437 0.7735 0.8795
0.4074 9.1034 528 0.7782 0.6437 0.7782 0.8822
0.4074 9.1379 530 0.7827 0.6488 0.7827 0.8847
0.4074 9.1724 532 0.7899 0.6391 0.7899 0.8888
0.4074 9.2069 534 0.7989 0.6323 0.7989 0.8938
0.4074 9.2414 536 0.8094 0.6215 0.8094 0.8997
0.4074 9.2759 538 0.8184 0.6145 0.8184 0.9046
0.4074 9.3103 540 0.8235 0.6167 0.8235 0.9074
0.4074 9.3448 542 0.8286 0.6167 0.8286 0.9103
0.4074 9.3793 544 0.8365 0.6173 0.8365 0.9146
0.4074 9.4138 546 0.8412 0.6078 0.8412 0.9172
0.4074 9.4483 548 0.8434 0.6078 0.8434 0.9184
0.4074 9.4828 550 0.8433 0.5952 0.8433 0.9183
0.4074 9.5172 552 0.8417 0.6046 0.8417 0.9174
0.4074 9.5517 554 0.8419 0.6078 0.8419 0.9175
0.4074 9.5862 556 0.8423 0.6078 0.8423 0.9178
0.4074 9.6207 558 0.8444 0.6182 0.8444 0.9189
0.4074 9.6552 560 0.8429 0.6182 0.8429 0.9181
0.4074 9.6897 562 0.8386 0.6182 0.8386 0.9157
0.4074 9.7241 564 0.8341 0.6182 0.8341 0.9133
0.4074 9.7586 566 0.8309 0.6182 0.8309 0.9116
0.4074 9.7931 568 0.8277 0.6091 0.8277 0.9098
0.4074 9.8276 570 0.8257 0.6091 0.8257 0.9087
0.4074 9.8621 572 0.8247 0.6091 0.8247 0.9081
0.4074 9.8966 574 0.8238 0.6187 0.8238 0.9076
0.4074 9.9310 576 0.8239 0.6091 0.8239 0.9077
0.4074 9.9655 578 0.8242 0.6091 0.8242 0.9079
0.4074 10.0 580 0.8242 0.6091 0.8242 0.9079

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
3
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run3_AugV5_k9_task1_organization

Finetuned
(4023)
this model