Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6898
  • Qwk: 0.6360
  • Mse: 0.6898
  • Rmse: 0.8306

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0202 2 4.0402 -0.0231 4.0402 2.0100
No log 0.0404 4 3.0226 0.0692 3.0226 1.7386
No log 0.0606 6 2.0617 0.1058 2.0617 1.4359
No log 0.0808 8 1.1022 0.1929 1.1022 1.0498
No log 0.1010 10 1.2230 0.0120 1.2230 1.1059
No log 0.1212 12 1.7696 -0.2639 1.7696 1.3303
No log 0.1414 14 1.9932 -0.0869 1.9932 1.4118
No log 0.1616 16 1.3918 -0.0522 1.3918 1.1797
No log 0.1818 18 1.2111 0.1616 1.2111 1.1005
No log 0.2020 20 1.2560 0.1811 1.2560 1.1207
No log 0.2222 22 1.0318 0.2798 1.0318 1.0158
No log 0.2424 24 0.9611 0.3560 0.9611 0.9804
No log 0.2626 26 0.9561 0.3663 0.9561 0.9778
No log 0.2828 28 0.9200 0.4106 0.9200 0.9592
No log 0.3030 30 0.8172 0.4470 0.8172 0.9040
No log 0.3232 32 0.7922 0.4635 0.7922 0.8901
No log 0.3434 34 0.8234 0.4748 0.8234 0.9074
No log 0.3636 36 1.0293 0.4391 1.0293 1.0146
No log 0.3838 38 1.1250 0.4203 1.1250 1.0607
No log 0.4040 40 1.0591 0.3906 1.0591 1.0291
No log 0.4242 42 0.8724 0.4459 0.8724 0.9340
No log 0.4444 44 0.7886 0.5360 0.7886 0.8880
No log 0.4646 46 0.7855 0.5354 0.7855 0.8863
No log 0.4848 48 0.8765 0.3957 0.8765 0.9362
No log 0.5051 50 0.8846 0.3959 0.8846 0.9405
No log 0.5253 52 0.8334 0.4051 0.8334 0.9129
No log 0.5455 54 0.7773 0.4799 0.7773 0.8816
No log 0.5657 56 0.8110 0.4037 0.8110 0.9006
No log 0.5859 58 0.9395 0.3326 0.9395 0.9693
No log 0.6061 60 0.8725 0.3700 0.8725 0.9341
No log 0.6263 62 0.7622 0.4630 0.7622 0.8731
No log 0.6465 64 0.8046 0.4498 0.8046 0.8970
No log 0.6667 66 0.8282 0.4592 0.8282 0.9100
No log 0.6869 68 0.8390 0.4900 0.8390 0.9160
No log 0.7071 70 0.8249 0.5384 0.8249 0.9083
No log 0.7273 72 0.8262 0.5376 0.8262 0.9090
No log 0.7475 74 0.7405 0.6140 0.7405 0.8605
No log 0.7677 76 0.7168 0.6312 0.7168 0.8466
No log 0.7879 78 0.7348 0.6098 0.7348 0.8572
No log 0.8081 80 0.8223 0.4280 0.8223 0.9068
No log 0.8283 82 0.9091 0.3420 0.9091 0.9535
No log 0.8485 84 0.7518 0.5131 0.7518 0.8671
No log 0.8687 86 0.6500 0.6553 0.6500 0.8063
No log 0.8889 88 0.6924 0.6523 0.6924 0.8321
No log 0.9091 90 0.7550 0.6176 0.7550 0.8689
No log 0.9293 92 0.7450 0.6427 0.7450 0.8631
No log 0.9495 94 0.8777 0.5760 0.8777 0.9368
No log 0.9697 96 0.8970 0.5686 0.8970 0.9471
No log 0.9899 98 0.8661 0.5742 0.8661 0.9306
No log 1.0101 100 0.9210 0.5555 0.9210 0.9597
No log 1.0303 102 0.9670 0.5429 0.9670 0.9834
No log 1.0505 104 0.9172 0.5108 0.9172 0.9577
No log 1.0707 106 0.7694 0.6265 0.7694 0.8772
No log 1.0909 108 0.8451 0.6268 0.8451 0.9193
No log 1.1111 110 0.8826 0.6121 0.8826 0.9395
No log 1.1313 112 0.7331 0.6219 0.7331 0.8562
No log 1.1515 114 0.7869 0.5242 0.7869 0.8871
No log 1.1717 116 0.9381 0.4088 0.9381 0.9685
No log 1.1919 118 0.9755 0.3809 0.9755 0.9877
No log 1.2121 120 0.8072 0.5411 0.8072 0.8984
No log 1.2323 122 0.6656 0.6413 0.6656 0.8159
No log 1.2525 124 0.6862 0.6200 0.6862 0.8284
No log 1.2727 126 0.6851 0.6346 0.6851 0.8277
No log 1.2929 128 0.6841 0.6326 0.6841 0.8271
No log 1.3131 130 0.7088 0.6493 0.7088 0.8419
No log 1.3333 132 0.7314 0.6406 0.7314 0.8552
No log 1.3535 134 0.8170 0.5874 0.8170 0.9039
No log 1.3737 136 0.8770 0.5362 0.8770 0.9365
No log 1.3939 138 0.7522 0.5747 0.7522 0.8673
No log 1.4141 140 0.6724 0.6311 0.6724 0.8200
No log 1.4343 142 0.6795 0.6264 0.6795 0.8243
No log 1.4545 144 0.6952 0.6221 0.6952 0.8338
No log 1.4747 146 0.7771 0.5991 0.7771 0.8815
No log 1.4949 148 0.7796 0.6103 0.7796 0.8829
No log 1.5152 150 0.7817 0.6020 0.7817 0.8841
No log 1.5354 152 0.8374 0.5444 0.8374 0.9151
No log 1.5556 154 0.9028 0.4827 0.9028 0.9501
No log 1.5758 156 0.8589 0.5388 0.8589 0.9267
No log 1.5960 158 0.8748 0.5037 0.8748 0.9353
No log 1.6162 160 0.8409 0.5268 0.8409 0.9170
No log 1.6364 162 0.7027 0.6199 0.7027 0.8383
No log 1.6566 164 0.6613 0.6589 0.6613 0.8132
No log 1.6768 166 0.6752 0.6241 0.6752 0.8217
No log 1.6970 168 0.7164 0.6456 0.7164 0.8464
No log 1.7172 170 0.8681 0.5877 0.8681 0.9317
No log 1.7374 172 0.9006 0.5697 0.9006 0.9490
No log 1.7576 174 0.9799 0.5615 0.9799 0.9899
No log 1.7778 176 0.9103 0.5739 0.9103 0.9541
No log 1.7980 178 0.8368 0.5595 0.8368 0.9148
No log 1.8182 180 0.8088 0.5648 0.8088 0.8993
No log 1.8384 182 0.7855 0.5770 0.7855 0.8863
No log 1.8586 184 0.7957 0.5629 0.7957 0.8920
No log 1.8788 186 0.6964 0.5870 0.6964 0.8345
No log 1.8990 188 0.6820 0.6065 0.6820 0.8258
No log 1.9192 190 0.6886 0.6152 0.6886 0.8298
No log 1.9394 192 0.6589 0.6170 0.6589 0.8117
No log 1.9596 194 0.6781 0.6085 0.6781 0.8235
No log 1.9798 196 0.6997 0.5506 0.6997 0.8365
No log 2.0 198 0.7137 0.5101 0.7137 0.8448
No log 2.0202 200 0.6937 0.5711 0.6937 0.8329
No log 2.0404 202 0.6857 0.5738 0.6857 0.8280
No log 2.0606 204 0.6932 0.5653 0.6932 0.8326
No log 2.0808 206 0.7837 0.5701 0.7837 0.8853
No log 2.1010 208 0.9936 0.5183 0.9936 0.9968
No log 2.1212 210 1.1052 0.4911 1.1052 1.0513
No log 2.1414 212 1.0218 0.5858 1.0218 1.0108
No log 2.1616 214 0.9456 0.6383 0.9456 0.9724
No log 2.1818 216 0.8564 0.6470 0.8564 0.9254
No log 2.2020 218 0.8028 0.5921 0.8028 0.8960
No log 2.2222 220 0.8974 0.5218 0.8974 0.9473
No log 2.2424 222 0.9091 0.4809 0.9091 0.9535
No log 2.2626 224 0.8667 0.4948 0.8667 0.9310
No log 2.2828 226 0.7853 0.5218 0.7853 0.8862
No log 2.3030 228 0.7420 0.5634 0.7420 0.8614
No log 2.3232 230 0.6370 0.6423 0.6370 0.7981
No log 2.3434 232 0.6376 0.6678 0.6376 0.7985
No log 2.3636 234 0.6399 0.6699 0.6399 0.8000
No log 2.3838 236 0.6259 0.6708 0.6259 0.7911
No log 2.4040 238 0.6869 0.6344 0.6869 0.8288
No log 2.4242 240 0.7629 0.6296 0.7629 0.8734
No log 2.4444 242 0.7509 0.6183 0.7509 0.8665
No log 2.4646 244 0.9028 0.5789 0.9028 0.9502
No log 2.4848 246 0.9107 0.5775 0.9107 0.9543
No log 2.5051 248 0.7457 0.6237 0.7457 0.8635
No log 2.5253 250 0.6504 0.6310 0.6504 0.8065
No log 2.5455 252 0.6950 0.5986 0.6950 0.8336
No log 2.5657 254 0.7660 0.5849 0.7660 0.8752
No log 2.5859 256 0.7147 0.5932 0.7147 0.8454
No log 2.6061 258 0.6487 0.6237 0.6487 0.8054
No log 2.6263 260 0.7919 0.5931 0.7919 0.8899
No log 2.6465 262 0.8791 0.5729 0.8791 0.9376
No log 2.6667 264 0.7583 0.6023 0.7583 0.8708
No log 2.6869 266 0.6892 0.6412 0.6892 0.8302
No log 2.7071 268 0.7018 0.6481 0.7018 0.8378
No log 2.7273 270 0.7426 0.6309 0.7426 0.8617
No log 2.7475 272 0.7165 0.6426 0.7165 0.8465
No log 2.7677 274 0.6954 0.6263 0.6955 0.8339
No log 2.7879 276 0.6827 0.5995 0.6827 0.8263
No log 2.8081 278 0.6839 0.5651 0.6839 0.8270
No log 2.8283 280 0.6939 0.5393 0.6939 0.8330
No log 2.8485 282 0.6648 0.5671 0.6648 0.8153
No log 2.8687 284 0.6398 0.6167 0.6398 0.7999
No log 2.8889 286 0.6855 0.5996 0.6855 0.8280
No log 2.9091 288 0.6661 0.6292 0.6661 0.8162
No log 2.9293 290 0.6133 0.6601 0.6133 0.7831
No log 2.9495 292 0.6515 0.6298 0.6515 0.8071
No log 2.9697 294 0.8307 0.5705 0.8307 0.9114
No log 2.9899 296 0.9326 0.5192 0.9326 0.9657
No log 3.0101 298 1.1018 0.4280 1.1018 1.0497
No log 3.0303 300 1.0019 0.5174 1.0019 1.0010
No log 3.0505 302 0.7944 0.5947 0.7944 0.8913
No log 3.0707 304 0.6969 0.6379 0.6969 0.8348
No log 3.0909 306 0.6640 0.6669 0.6640 0.8149
No log 3.1111 308 0.7964 0.6211 0.7964 0.8924
No log 3.1313 310 0.8303 0.6307 0.8303 0.9112
No log 3.1515 312 0.7130 0.6497 0.7130 0.8444
No log 3.1717 314 0.6433 0.6740 0.6433 0.8021
No log 3.1919 316 0.6965 0.6303 0.6965 0.8346
No log 3.2121 318 0.7223 0.5907 0.7223 0.8499
No log 3.2323 320 0.7514 0.5776 0.7514 0.8668
No log 3.2525 322 0.6793 0.6022 0.6793 0.8242
No log 3.2727 324 0.6307 0.6541 0.6307 0.7942
No log 3.2929 326 0.6780 0.6659 0.6780 0.8234
No log 3.3131 328 0.6590 0.6759 0.6590 0.8118
No log 3.3333 330 0.6941 0.6333 0.6941 0.8331
No log 3.3535 332 0.9524 0.5317 0.9524 0.9759
No log 3.3737 334 1.1053 0.4894 1.1053 1.0513
No log 3.3939 336 1.0201 0.5250 1.0201 1.0100
No log 3.4141 338 0.8433 0.5750 0.8433 0.9183
No log 3.4343 340 0.7088 0.6252 0.7088 0.8419
No log 3.4545 342 0.6537 0.6297 0.6537 0.8085
No log 3.4747 344 0.6326 0.6473 0.6326 0.7954
No log 3.4949 346 0.6334 0.6421 0.6334 0.7958
No log 3.5152 348 0.6640 0.6335 0.6640 0.8149
No log 3.5354 350 0.7642 0.6073 0.7642 0.8742
No log 3.5556 352 0.7997 0.6126 0.7997 0.8943
No log 3.5758 354 0.7741 0.6287 0.7741 0.8798
No log 3.5960 356 0.8095 0.6078 0.8095 0.8997
No log 3.6162 358 0.8985 0.5974 0.8985 0.9479
No log 3.6364 360 0.7966 0.6334 0.7966 0.8925
No log 3.6566 362 0.7220 0.6614 0.7220 0.8497
No log 3.6768 364 0.7522 0.6703 0.7522 0.8673
No log 3.6970 366 0.7093 0.6860 0.7093 0.8422
No log 3.7172 368 0.6952 0.6572 0.6952 0.8338
No log 3.7374 370 0.8273 0.5971 0.8273 0.9096
No log 3.7576 372 0.8287 0.5917 0.8287 0.9103
No log 3.7778 374 0.6975 0.6184 0.6975 0.8351
No log 3.7980 376 0.6474 0.6348 0.6474 0.8046
No log 3.8182 378 0.6271 0.6371 0.6271 0.7919
No log 3.8384 380 0.6239 0.6900 0.6239 0.7899
No log 3.8586 382 0.6775 0.6670 0.6775 0.8231
No log 3.8788 384 0.8260 0.6463 0.8260 0.9088
No log 3.8990 386 0.8463 0.6271 0.8463 0.9200
No log 3.9192 388 0.7470 0.6534 0.7470 0.8643
No log 3.9394 390 0.6208 0.6872 0.6208 0.7879
No log 3.9596 392 0.5923 0.6876 0.5923 0.7696
No log 3.9798 394 0.5787 0.6786 0.5787 0.7607
No log 4.0 396 0.5896 0.6485 0.5896 0.7679
No log 4.0202 398 0.6086 0.6378 0.6086 0.7801
No log 4.0404 400 0.6522 0.6083 0.6522 0.8076
No log 4.0606 402 0.6528 0.6246 0.6528 0.8079
No log 4.0808 404 0.6374 0.6375 0.6374 0.7984
No log 4.1010 406 0.6177 0.6757 0.6177 0.7859
No log 4.1212 408 0.6265 0.6734 0.6265 0.7915
No log 4.1414 410 0.6625 0.6623 0.6625 0.8139
No log 4.1616 412 0.7296 0.6480 0.7296 0.8542
No log 4.1818 414 0.7049 0.6418 0.7049 0.8396
No log 4.2020 416 0.7323 0.6363 0.7323 0.8557
No log 4.2222 418 0.6663 0.6552 0.6663 0.8163
No log 4.2424 420 0.6213 0.6515 0.6213 0.7882
No log 4.2626 422 0.5995 0.6585 0.5995 0.7742
No log 4.2828 424 0.6429 0.6575 0.6429 0.8018
No log 4.3030 426 0.7871 0.6092 0.7871 0.8872
No log 4.3232 428 0.8127 0.6192 0.8127 0.9015
No log 4.3434 430 0.7159 0.6693 0.7159 0.8461
No log 4.3636 432 0.6279 0.6975 0.6279 0.7924
No log 4.3838 434 0.6101 0.6845 0.6101 0.7811
No log 4.4040 436 0.6074 0.6554 0.6074 0.7794
No log 4.4242 438 0.7050 0.6366 0.7050 0.8396
No log 4.4444 440 0.7419 0.5829 0.7419 0.8613
No log 4.4646 442 0.6782 0.5986 0.6782 0.8235
No log 4.4848 444 0.6110 0.6277 0.6110 0.7816
No log 4.5051 446 0.6721 0.5789 0.6721 0.8198
No log 4.5253 448 0.6792 0.5965 0.6792 0.8241
No log 4.5455 450 0.6322 0.6680 0.6322 0.7951
No log 4.5657 452 0.7133 0.6338 0.7133 0.8446
No log 4.5859 454 0.9591 0.5798 0.9591 0.9793
No log 4.6061 456 1.0211 0.5587 1.0211 1.0105
No log 4.6263 458 0.8511 0.6297 0.8511 0.9225
No log 4.6465 460 0.6945 0.6405 0.6945 0.8334
No log 4.6667 462 0.6997 0.6557 0.6997 0.8365
No log 4.6869 464 0.6716 0.6353 0.6716 0.8195
No log 4.7071 466 0.6634 0.6514 0.6634 0.8145
No log 4.7273 468 0.7307 0.6236 0.7307 0.8548
No log 4.7475 470 0.7557 0.6322 0.7557 0.8693
No log 4.7677 472 0.7302 0.6571 0.7302 0.8545
No log 4.7879 474 0.7985 0.6418 0.7985 0.8936
No log 4.8081 476 0.7822 0.6405 0.7822 0.8844
No log 4.8283 478 0.7447 0.6508 0.7447 0.8629
No log 4.8485 480 0.7655 0.6510 0.7655 0.8749
No log 4.8687 482 0.7683 0.6549 0.7683 0.8765
No log 4.8889 484 0.8264 0.6174 0.8264 0.9090
No log 4.9091 486 0.7629 0.6231 0.7629 0.8735
No log 4.9293 488 0.7017 0.6311 0.7017 0.8377
No log 4.9495 490 0.6641 0.6634 0.6641 0.8149
No log 4.9697 492 0.7021 0.6112 0.7021 0.8379
No log 4.9899 494 0.8478 0.5978 0.8478 0.9208
No log 5.0101 496 0.9165 0.5812 0.9165 0.9574
No log 5.0303 498 0.7967 0.6511 0.7967 0.8926
0.5907 5.0505 500 0.7407 0.6459 0.7407 0.8607
0.5907 5.0707 502 0.7590 0.6400 0.7590 0.8712
0.5907 5.0909 504 0.8666 0.6516 0.8666 0.9309
0.5907 5.1111 506 1.1782 0.5243 1.1782 1.0855
0.5907 5.1313 508 1.3530 0.4584 1.3530 1.1632
0.5907 5.1515 510 1.2073 0.5181 1.2073 1.0988
0.5907 5.1717 512 0.8571 0.5859 0.8571 0.9258
0.5907 5.1919 514 0.6898 0.6360 0.6898 0.8306

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask7_organization

Finetuned
(4019)
this model