ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k9_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7526
  • Qwk: 0.7037
  • Mse: 0.7526
  • Rmse: 0.8675

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0392 2 5.1131 -0.0110 5.1131 2.2612
No log 0.0784 4 2.9579 0.0845 2.9579 1.7199
No log 0.1176 6 1.5513 0.1530 1.5513 1.2455
No log 0.1569 8 1.3208 0.2337 1.3208 1.1493
No log 0.1961 10 1.1760 0.1981 1.1760 1.0845
No log 0.2353 12 1.1525 0.1680 1.1525 1.0735
No log 0.2745 14 1.2286 0.1963 1.2286 1.1084
No log 0.3137 16 1.2379 0.2169 1.2379 1.1126
No log 0.3529 18 1.3230 0.1970 1.3230 1.1502
No log 0.3922 20 1.3232 0.2038 1.3232 1.1503
No log 0.4314 22 1.2445 0.3956 1.2445 1.1156
No log 0.4706 24 1.2228 0.3895 1.2228 1.1058
No log 0.5098 26 1.1993 0.3555 1.1993 1.0951
No log 0.5490 28 1.1523 0.3697 1.1523 1.0734
No log 0.5882 30 1.1222 0.2821 1.1222 1.0594
No log 0.6275 32 1.1143 0.2161 1.1143 1.0556
No log 0.6667 34 1.0876 0.2361 1.0876 1.0429
No log 0.7059 36 1.0826 0.2068 1.0826 1.0405
No log 0.7451 38 1.0462 0.3439 1.0462 1.0229
No log 0.7843 40 0.9317 0.4057 0.9317 0.9653
No log 0.8235 42 0.8296 0.4985 0.8296 0.9108
No log 0.8627 44 0.9072 0.4469 0.9072 0.9525
No log 0.9020 46 0.9031 0.4450 0.9031 0.9503
No log 0.9412 48 0.8222 0.4622 0.8222 0.9068
No log 0.9804 50 0.7772 0.5118 0.7772 0.8816
No log 1.0196 52 0.7140 0.5570 0.7140 0.8450
No log 1.0588 54 0.6815 0.6102 0.6815 0.8255
No log 1.0980 56 0.6923 0.6274 0.6923 0.8320
No log 1.1373 58 0.6510 0.6632 0.6510 0.8069
No log 1.1765 60 0.7052 0.6291 0.7052 0.8397
No log 1.2157 62 0.7408 0.6600 0.7408 0.8607
No log 1.2549 64 0.6490 0.7079 0.6490 0.8056
No log 1.2941 66 0.6000 0.7005 0.6000 0.7746
No log 1.3333 68 0.5767 0.7073 0.5767 0.7594
No log 1.3725 70 0.5483 0.7152 0.5483 0.7405
No log 1.4118 72 0.5860 0.6762 0.5860 0.7655
No log 1.4510 74 0.6540 0.6536 0.6540 0.8087
No log 1.4902 76 0.6133 0.7035 0.6133 0.7831
No log 1.5294 78 0.5758 0.7047 0.5758 0.7588
No log 1.5686 80 0.7008 0.6809 0.7008 0.8372
No log 1.6078 82 0.7746 0.6634 0.7746 0.8801
No log 1.6471 84 0.8371 0.6325 0.8371 0.9149
No log 1.6863 86 0.7777 0.6482 0.7777 0.8819
No log 1.7255 88 0.6462 0.6388 0.6462 0.8038
No log 1.7647 90 0.5828 0.6574 0.5828 0.7634
No log 1.8039 92 0.5788 0.6708 0.5788 0.7608
No log 1.8431 94 0.6208 0.6893 0.6208 0.7879
No log 1.8824 96 0.7139 0.6939 0.7139 0.8449
No log 1.9216 98 0.8328 0.6579 0.8328 0.9126
No log 1.9608 100 0.7470 0.6910 0.7470 0.8643
No log 2.0 102 0.6565 0.7241 0.6565 0.8103
No log 2.0392 104 0.6310 0.7185 0.6310 0.7944
No log 2.0784 106 0.6366 0.7164 0.6366 0.7979
No log 2.1176 108 0.7074 0.7181 0.7074 0.8411
No log 2.1569 110 0.7636 0.6730 0.7636 0.8738
No log 2.1961 112 0.8153 0.6627 0.8153 0.9030
No log 2.2353 114 0.8207 0.6748 0.8207 0.9059
No log 2.2745 116 0.7831 0.6732 0.7831 0.8850
No log 2.3137 118 0.8418 0.6647 0.8418 0.9175
No log 2.3529 120 1.0512 0.6109 1.0512 1.0253
No log 2.3922 122 1.2573 0.5668 1.2573 1.1213
No log 2.4314 124 1.1272 0.5607 1.1272 1.0617
No log 2.4706 126 0.8736 0.6357 0.8736 0.9347
No log 2.5098 128 0.7222 0.6671 0.7222 0.8498
No log 2.5490 130 0.6292 0.7404 0.6292 0.7932
No log 2.5882 132 0.6237 0.7562 0.6237 0.7897
No log 2.6275 134 0.6609 0.7311 0.6609 0.8130
No log 2.6667 136 0.7988 0.6812 0.7988 0.8938
No log 2.7059 138 0.9094 0.6612 0.9094 0.9536
No log 2.7451 140 0.8797 0.6595 0.8797 0.9379
No log 2.7843 142 0.7233 0.6788 0.7233 0.8505
No log 2.8235 144 0.6688 0.6908 0.6688 0.8178
No log 2.8627 146 0.6864 0.7135 0.6864 0.8285
No log 2.9020 148 0.7704 0.6317 0.7704 0.8777
No log 2.9412 150 0.8465 0.6233 0.8465 0.9200
No log 2.9804 152 0.8296 0.6521 0.8296 0.9108
No log 3.0196 154 0.9163 0.6091 0.9163 0.9572
No log 3.0588 156 0.9063 0.6345 0.9063 0.9520
No log 3.0980 158 0.8625 0.6207 0.8625 0.9287
No log 3.1373 160 0.7881 0.6723 0.7881 0.8878
No log 3.1765 162 0.7832 0.6607 0.7832 0.8850
No log 3.2157 164 0.8960 0.6233 0.8960 0.9466
No log 3.2549 166 0.9768 0.6146 0.9768 0.9883
No log 3.2941 168 0.9792 0.6051 0.9792 0.9896
No log 3.3333 170 0.8899 0.6228 0.8899 0.9433
No log 3.3725 172 0.8737 0.6273 0.8737 0.9347
No log 3.4118 174 0.8365 0.6235 0.8365 0.9146
No log 3.4510 176 0.9055 0.6134 0.9055 0.9516
No log 3.4902 178 0.8792 0.5935 0.8792 0.9377
No log 3.5294 180 0.7653 0.6436 0.7653 0.8748
No log 3.5686 182 0.6811 0.7300 0.6811 0.8253
No log 3.6078 184 0.6761 0.7230 0.6761 0.8223
No log 3.6471 186 0.6979 0.7027 0.6979 0.8354
No log 3.6863 188 0.8207 0.6294 0.8207 0.9059
No log 3.7255 190 0.9821 0.5904 0.9821 0.9910
No log 3.7647 192 0.9524 0.6025 0.9524 0.9759
No log 3.8039 194 0.8170 0.6857 0.8170 0.9039
No log 3.8431 196 0.7873 0.6863 0.7873 0.8873
No log 3.8824 198 0.8258 0.6787 0.8258 0.9087
No log 3.9216 200 0.8717 0.7005 0.8717 0.9336
No log 3.9608 202 0.9039 0.6755 0.9039 0.9507
No log 4.0 204 0.9494 0.6572 0.9494 0.9744
No log 4.0392 206 1.0639 0.6109 1.0639 1.0315
No log 4.0784 208 1.0314 0.6179 1.0314 1.0156
No log 4.1176 210 0.9062 0.6268 0.9062 0.9519
No log 4.1569 212 0.8041 0.6809 0.8041 0.8967
No log 4.1961 214 0.7983 0.6753 0.7983 0.8935
No log 4.2353 216 0.8432 0.6643 0.8432 0.9182
No log 4.2745 218 0.9548 0.6732 0.9548 0.9771
No log 4.3137 220 0.9650 0.6590 0.9650 0.9823
No log 4.3529 222 1.0210 0.6445 1.0210 1.0104
No log 4.3922 224 1.0041 0.5805 1.0041 1.0020
No log 4.4314 226 0.9107 0.6055 0.9107 0.9543
No log 4.4706 228 0.8290 0.6685 0.8290 0.9105
No log 4.5098 230 0.8291 0.6669 0.8291 0.9105
No log 4.5490 232 0.8601 0.6411 0.8601 0.9274
No log 4.5882 234 0.7938 0.6510 0.7938 0.8910
No log 4.6275 236 0.6705 0.6925 0.6705 0.8188
No log 4.6667 238 0.6302 0.6876 0.6302 0.7938
No log 4.7059 240 0.6283 0.6946 0.6283 0.7927
No log 4.7451 242 0.6160 0.7053 0.6160 0.7849
No log 4.7843 244 0.6343 0.7330 0.6343 0.7964
No log 4.8235 246 0.7233 0.7183 0.7233 0.8505
No log 4.8627 248 0.9326 0.6438 0.9326 0.9657
No log 4.9020 250 1.0175 0.6317 1.0175 1.0087
No log 4.9412 252 0.9741 0.6290 0.9741 0.9870
No log 4.9804 254 0.8250 0.6450 0.8250 0.9083
No log 5.0196 256 0.6845 0.7407 0.6845 0.8273
No log 5.0588 258 0.6401 0.7330 0.6401 0.8001
No log 5.0980 260 0.6507 0.7324 0.6507 0.8067
No log 5.1373 262 0.6840 0.7138 0.6840 0.8270
No log 5.1765 264 0.7077 0.6988 0.7077 0.8413
No log 5.2157 266 0.7102 0.6906 0.7102 0.8427
No log 5.2549 268 0.6853 0.7213 0.6853 0.8278
No log 5.2941 270 0.6760 0.7404 0.6760 0.8222
No log 5.3333 272 0.6898 0.7282 0.6898 0.8305
No log 5.3725 274 0.7399 0.6879 0.7399 0.8602
No log 5.4118 276 0.8141 0.6652 0.8141 0.9023
No log 5.4510 278 0.8085 0.6652 0.8085 0.8991
No log 5.4902 280 0.8036 0.6730 0.8036 0.8964
No log 5.5294 282 0.8292 0.6872 0.8292 0.9106
No log 5.5686 284 0.8010 0.6817 0.8010 0.8950
No log 5.6078 286 0.7462 0.6938 0.7462 0.8638
No log 5.6471 288 0.6686 0.7095 0.6686 0.8177
No log 5.6863 290 0.6412 0.7216 0.6412 0.8007
No log 5.7255 292 0.6580 0.7210 0.6580 0.8112
No log 5.7647 294 0.6374 0.7335 0.6374 0.7984
No log 5.8039 296 0.6114 0.7158 0.6114 0.7819
No log 5.8431 298 0.6485 0.7218 0.6485 0.8053
No log 5.8824 300 0.7208 0.7084 0.7208 0.8490
No log 5.9216 302 0.8217 0.7031 0.8217 0.9065
No log 5.9608 304 0.8504 0.6967 0.8504 0.9222
No log 6.0 306 0.8372 0.6869 0.8372 0.9150
No log 6.0392 308 0.7720 0.6898 0.7720 0.8786
No log 6.0784 310 0.6975 0.7039 0.6975 0.8351
No log 6.1176 312 0.6802 0.6976 0.6802 0.8247
No log 6.1569 314 0.6906 0.7046 0.6906 0.8310
No log 6.1961 316 0.7091 0.6994 0.7091 0.8421
No log 6.2353 318 0.7266 0.6972 0.7266 0.8524
No log 6.2745 320 0.7296 0.6955 0.7296 0.8542
No log 6.3137 322 0.7211 0.6972 0.7211 0.8492
No log 6.3529 324 0.7066 0.6956 0.7066 0.8406
No log 6.3922 326 0.7345 0.6917 0.7345 0.8570
No log 6.4314 328 0.7494 0.6878 0.7494 0.8657
No log 6.4706 330 0.7741 0.6750 0.7741 0.8799
No log 6.5098 332 0.7718 0.6756 0.7718 0.8785
No log 6.5490 334 0.7434 0.6889 0.7434 0.8622
No log 6.5882 336 0.6888 0.7315 0.6888 0.8299
No log 6.6275 338 0.6914 0.7320 0.6914 0.8315
No log 6.6667 340 0.7412 0.7109 0.7412 0.8609
No log 6.7059 342 0.8469 0.6762 0.8469 0.9203
No log 6.7451 344 0.9056 0.6817 0.9056 0.9516
No log 6.7843 346 0.8719 0.6734 0.8719 0.9338
No log 6.8235 348 0.8152 0.6803 0.8152 0.9029
No log 6.8627 350 0.7667 0.6834 0.7667 0.8756
No log 6.9020 352 0.7345 0.7274 0.7345 0.8570
No log 6.9412 354 0.6934 0.7462 0.6934 0.8327
No log 6.9804 356 0.6594 0.7469 0.6594 0.8121
No log 7.0196 358 0.6533 0.7445 0.6533 0.8082
No log 7.0588 360 0.6879 0.7335 0.6879 0.8294
No log 7.0980 362 0.7874 0.6942 0.7874 0.8873
No log 7.1373 364 0.8635 0.6805 0.8635 0.9293
No log 7.1765 366 0.8426 0.6914 0.8426 0.9179
No log 7.2157 368 0.7738 0.7004 0.7738 0.8797
No log 7.2549 370 0.7232 0.6944 0.7232 0.8504
No log 7.2941 372 0.6911 0.7461 0.6911 0.8313
No log 7.3333 374 0.7109 0.7302 0.7109 0.8432
No log 7.3725 376 0.7690 0.7004 0.7690 0.8770
No log 7.4118 378 0.8617 0.6729 0.8617 0.9283
No log 7.4510 380 0.9384 0.6745 0.9384 0.9687
No log 7.4902 382 0.9414 0.6745 0.9414 0.9703
No log 7.5294 384 0.9014 0.6729 0.9014 0.9494
No log 7.5686 386 0.8296 0.6671 0.8296 0.9108
No log 7.6078 388 0.7782 0.6926 0.7782 0.8822
No log 7.6471 390 0.7505 0.6937 0.7505 0.8663
No log 7.6863 392 0.7475 0.6976 0.7475 0.8646
No log 7.7255 394 0.7425 0.7163 0.7425 0.8617
No log 7.7647 396 0.7441 0.7010 0.7441 0.8626
No log 7.8039 398 0.7404 0.7170 0.7404 0.8605
No log 7.8431 400 0.7565 0.6905 0.7565 0.8698
No log 7.8824 402 0.7769 0.6833 0.7769 0.8814
No log 7.9216 404 0.7628 0.6816 0.7628 0.8734
No log 7.9608 406 0.7570 0.6904 0.7570 0.8700
No log 8.0 408 0.7544 0.6904 0.7544 0.8686
No log 8.0392 410 0.7680 0.6833 0.7680 0.8763
No log 8.0784 412 0.8007 0.6794 0.8007 0.8948
No log 8.1176 414 0.8065 0.6794 0.8065 0.8980
No log 8.1569 416 0.8444 0.6654 0.8444 0.9189
No log 8.1961 418 0.8622 0.6555 0.8622 0.9285
No log 8.2353 420 0.8460 0.6654 0.8460 0.9198
No log 8.2745 422 0.8003 0.6794 0.8003 0.8946
No log 8.3137 424 0.7632 0.6965 0.7632 0.8736
No log 8.3529 426 0.7412 0.7089 0.7412 0.8610
No log 8.3922 428 0.7395 0.7166 0.7395 0.8600
No log 8.4314 430 0.7389 0.7056 0.7389 0.8596
No log 8.4706 432 0.7610 0.7004 0.7610 0.8723
No log 8.5098 434 0.8020 0.6932 0.8020 0.8955
No log 8.5490 436 0.8257 0.6931 0.8257 0.9087
No log 8.5882 438 0.8414 0.6845 0.8414 0.9173
No log 8.6275 440 0.8336 0.6807 0.8336 0.9130
No log 8.6667 442 0.8315 0.6807 0.8315 0.9119
No log 8.7059 444 0.8145 0.6794 0.8145 0.9025
No log 8.7451 446 0.7813 0.6882 0.7813 0.8839
No log 8.7843 448 0.7449 0.7155 0.7449 0.8631
No log 8.8235 450 0.7322 0.7194 0.7322 0.8557
No log 8.8627 452 0.7264 0.7139 0.7264 0.8523
No log 8.9020 454 0.7122 0.7308 0.7122 0.8439
No log 8.9412 456 0.7060 0.7308 0.7060 0.8402
No log 8.9804 458 0.7174 0.7147 0.7174 0.8470
No log 9.0196 460 0.7395 0.7147 0.7395 0.8599
No log 9.0588 462 0.7733 0.6970 0.7733 0.8794
No log 9.0980 464 0.7974 0.6794 0.7974 0.8930
No log 9.1373 466 0.8123 0.6871 0.8123 0.9013
No log 9.1765 468 0.8166 0.6883 0.8166 0.9037
No log 9.2157 470 0.8196 0.6883 0.8196 0.9053
No log 9.2549 472 0.8161 0.6883 0.8161 0.9034
No log 9.2941 474 0.8035 0.6883 0.8035 0.8964
No log 9.3333 476 0.7951 0.6845 0.7951 0.8917
No log 9.3725 478 0.7894 0.6845 0.7894 0.8885
No log 9.4118 480 0.7885 0.6845 0.7885 0.8880
No log 9.4510 482 0.7934 0.6845 0.7934 0.8907
No log 9.4902 484 0.7952 0.6845 0.7952 0.8917
No log 9.5294 486 0.7867 0.6845 0.7867 0.8870
No log 9.5686 488 0.7735 0.6932 0.7735 0.8795
No log 9.6078 490 0.7616 0.7019 0.7616 0.8727
No log 9.6471 492 0.7522 0.7037 0.7522 0.8673
No log 9.6863 494 0.7472 0.7072 0.7472 0.8644
No log 9.7255 496 0.7445 0.7072 0.7445 0.8629
No log 9.7647 498 0.7469 0.7054 0.7469 0.8642
0.4262 9.8039 500 0.7484 0.7054 0.7484 0.8651
0.4262 9.8431 502 0.7499 0.7054 0.7499 0.8660
0.4262 9.8824 504 0.7504 0.7054 0.7504 0.8663
0.4262 9.9216 506 0.7516 0.7054 0.7516 0.8669
0.4262 9.9608 508 0.7522 0.7037 0.7522 0.8673
0.4262 10.0 510 0.7526 0.7037 0.7526 0.8675

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k9_task1_organization

Finetuned
(4023)
this model