Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask1_vocabulary

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6707
  • Qwk: 0.5877
  • Mse: 0.6707
  • Rmse: 0.8190

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0213 2 4.7716 0.0061 4.7716 2.1844
No log 0.0426 4 3.4176 0.0642 3.4176 1.8487
No log 0.0638 6 1.7463 0.0613 1.7463 1.3215
No log 0.0851 8 1.0577 0.1272 1.0577 1.0285
No log 0.1064 10 0.8767 0.0681 0.8767 0.9363
No log 0.1277 12 0.8856 0.0163 0.8856 0.9410
No log 0.1489 14 0.8603 0.1925 0.8603 0.9275
No log 0.1702 16 0.7581 0.2550 0.7581 0.8707
No log 0.1915 18 0.7899 0.2904 0.7899 0.8888
No log 0.2128 20 0.7188 0.2600 0.7188 0.8478
No log 0.2340 22 0.6563 0.3695 0.6563 0.8101
No log 0.2553 24 0.6800 0.4289 0.6800 0.8246
No log 0.2766 26 0.6434 0.4623 0.6434 0.8021
No log 0.2979 28 0.5992 0.4024 0.5992 0.7741
No log 0.3191 30 0.6116 0.3470 0.6116 0.7820
No log 0.3404 32 0.6149 0.3969 0.6149 0.7842
No log 0.3617 34 0.5521 0.4691 0.5521 0.7430
No log 0.3830 36 0.6994 0.4862 0.6994 0.8363
No log 0.4043 38 0.7922 0.5029 0.7922 0.8901
No log 0.4255 40 0.6624 0.5139 0.6624 0.8139
No log 0.4468 42 0.5370 0.5638 0.5370 0.7328
No log 0.4681 44 0.5833 0.5196 0.5833 0.7638
No log 0.4894 46 0.6491 0.4356 0.6491 0.8056
No log 0.5106 48 0.6396 0.4795 0.6396 0.7998
No log 0.5319 50 0.5493 0.5539 0.5493 0.7411
No log 0.5532 52 0.5610 0.6330 0.5610 0.7490
No log 0.5745 54 0.5613 0.5191 0.5613 0.7492
No log 0.5957 56 0.5956 0.4969 0.5956 0.7718
No log 0.6170 58 0.6653 0.4589 0.6653 0.8157
No log 0.6383 60 0.6855 0.4498 0.6855 0.8280
No log 0.6596 62 0.7015 0.4511 0.7015 0.8375
No log 0.6809 64 0.6243 0.4537 0.6243 0.7901
No log 0.7021 66 0.5675 0.5476 0.5675 0.7533
No log 0.7234 68 0.5642 0.5976 0.5642 0.7511
No log 0.7447 70 0.5920 0.6045 0.5920 0.7694
No log 0.7660 72 0.5988 0.5822 0.5988 0.7738
No log 0.7872 74 0.6028 0.5872 0.6028 0.7764
No log 0.8085 76 0.5218 0.5139 0.5218 0.7224
No log 0.8298 78 0.5210 0.5310 0.5210 0.7218
No log 0.8511 80 0.5217 0.4901 0.5217 0.7223
No log 0.8723 82 0.5189 0.4856 0.5189 0.7203
No log 0.8936 84 0.6043 0.4811 0.6043 0.7774
No log 0.9149 86 0.7216 0.4858 0.7216 0.8495
No log 0.9362 88 0.6549 0.5043 0.6549 0.8093
No log 0.9574 90 0.5761 0.5735 0.5761 0.7590
No log 0.9787 92 0.5426 0.5965 0.5426 0.7366
No log 1.0 94 0.5143 0.5733 0.5143 0.7172
No log 1.0213 96 0.5158 0.6104 0.5158 0.7182
No log 1.0426 98 0.5344 0.5831 0.5344 0.7310
No log 1.0638 100 0.5019 0.5828 0.5019 0.7084
No log 1.0851 102 0.4923 0.5992 0.4923 0.7017
No log 1.1064 104 0.5280 0.5581 0.5280 0.7266
No log 1.1277 106 0.6018 0.4617 0.6018 0.7758
No log 1.1489 108 0.6414 0.4221 0.6414 0.8009
No log 1.1702 110 0.6469 0.4570 0.6469 0.8043
No log 1.1915 112 0.5485 0.5323 0.5485 0.7406
No log 1.2128 114 0.5689 0.5935 0.5689 0.7543
No log 1.2340 116 0.5362 0.6423 0.5362 0.7322
No log 1.2553 118 0.5473 0.5671 0.5473 0.7398
No log 1.2766 120 0.7222 0.4384 0.7222 0.8498
No log 1.2979 122 0.7842 0.3728 0.7842 0.8856
No log 1.3191 124 0.6842 0.4401 0.6842 0.8272
No log 1.3404 126 0.5328 0.5969 0.5328 0.7299
No log 1.3617 128 0.5080 0.6219 0.5080 0.7127
No log 1.3830 130 0.5044 0.6137 0.5044 0.7102
No log 1.4043 132 0.6019 0.5261 0.6019 0.7758
No log 1.4255 134 0.6758 0.4582 0.6758 0.8220
No log 1.4468 136 0.6587 0.4847 0.6587 0.8116
No log 1.4681 138 0.5721 0.4918 0.5721 0.7564
No log 1.4894 140 0.5246 0.5352 0.5246 0.7243
No log 1.5106 142 0.5140 0.5633 0.5140 0.7170
No log 1.5319 144 0.4976 0.5624 0.4976 0.7054
No log 1.5532 146 0.4759 0.5524 0.4759 0.6899
No log 1.5745 148 0.4959 0.5723 0.4959 0.7042
No log 1.5957 150 0.5241 0.5809 0.5241 0.7239
No log 1.6170 152 0.6275 0.5153 0.6275 0.7921
No log 1.6383 154 0.6567 0.5019 0.6567 0.8104
No log 1.6596 156 0.5710 0.5822 0.5710 0.7556
No log 1.6809 158 0.5899 0.6435 0.5899 0.7680
No log 1.7021 160 0.7633 0.5911 0.7633 0.8737
No log 1.7234 162 0.7968 0.5549 0.7968 0.8926
No log 1.7447 164 0.6056 0.5889 0.6056 0.7782
No log 1.7660 166 0.4716 0.6147 0.4716 0.6867
No log 1.7872 168 0.5586 0.4651 0.5586 0.7474
No log 1.8085 170 0.6903 0.3404 0.6903 0.8308
No log 1.8298 172 0.7038 0.3352 0.7038 0.8390
No log 1.8511 174 0.6381 0.4701 0.6381 0.7988
No log 1.8723 176 0.5344 0.5368 0.5344 0.7310
No log 1.8936 178 0.4968 0.5533 0.4968 0.7048
No log 1.9149 180 0.4936 0.5845 0.4936 0.7026
No log 1.9362 182 0.5394 0.6101 0.5394 0.7344
No log 1.9574 184 0.5900 0.5861 0.5900 0.7681
No log 1.9787 186 0.5856 0.6134 0.5856 0.7652
No log 2.0 188 0.5335 0.6314 0.5335 0.7304
No log 2.0213 190 0.5366 0.6201 0.5366 0.7325
No log 2.0426 192 0.5183 0.5672 0.5183 0.7200
No log 2.0638 194 0.5136 0.5783 0.5136 0.7166
No log 2.0851 196 0.5012 0.5835 0.5012 0.7080
No log 2.1064 198 0.4903 0.5846 0.4903 0.7002
No log 2.1277 200 0.5010 0.5795 0.5010 0.7078
No log 2.1489 202 0.4787 0.6120 0.4787 0.6919
No log 2.1702 204 0.4758 0.6155 0.4758 0.6898
No log 2.1915 206 0.4784 0.5994 0.4784 0.6917
No log 2.2128 208 0.4765 0.5642 0.4765 0.6903
No log 2.2340 210 0.4780 0.5461 0.4780 0.6914
No log 2.2553 212 0.4949 0.5748 0.4949 0.7035
No log 2.2766 214 0.4878 0.5223 0.4878 0.6984
No log 2.2979 216 0.4922 0.5546 0.4922 0.7016
No log 2.3191 218 0.4980 0.5928 0.4980 0.7057
No log 2.3404 220 0.5921 0.6073 0.5921 0.7695
No log 2.3617 222 0.6866 0.5840 0.6866 0.8286
No log 2.3830 224 0.6672 0.6017 0.6672 0.8168
No log 2.4043 226 0.6191 0.6104 0.6191 0.7869
No log 2.4255 228 0.6296 0.6115 0.6296 0.7935
No log 2.4468 230 0.6724 0.5196 0.6724 0.8200
No log 2.4681 232 0.6914 0.4334 0.6914 0.8315
No log 2.4894 234 0.6487 0.4212 0.6487 0.8054
No log 2.5106 236 0.5641 0.4620 0.5641 0.7511
No log 2.5319 238 0.5180 0.4945 0.5180 0.7197
No log 2.5532 240 0.5095 0.5406 0.5095 0.7138
No log 2.5745 242 0.5203 0.5520 0.5203 0.7213
No log 2.5957 244 0.5425 0.6137 0.5425 0.7366
No log 2.6170 246 0.5984 0.6209 0.5984 0.7736
No log 2.6383 248 0.6131 0.6065 0.6131 0.7830
No log 2.6596 250 0.5761 0.6206 0.5761 0.7590
No log 2.6809 252 0.5368 0.5982 0.5368 0.7327
No log 2.7021 254 0.5099 0.5644 0.5099 0.7141
No log 2.7234 256 0.4996 0.5601 0.4996 0.7069
No log 2.7447 258 0.4970 0.5458 0.4970 0.7050
No log 2.7660 260 0.5030 0.5160 0.5030 0.7093
No log 2.7872 262 0.5239 0.5190 0.5239 0.7238
No log 2.8085 264 0.5471 0.4927 0.5471 0.7396
No log 2.8298 266 0.5690 0.5584 0.5690 0.7543
No log 2.8511 268 0.5465 0.6024 0.5465 0.7393
No log 2.8723 270 0.5055 0.6450 0.5055 0.7110
No log 2.8936 272 0.5277 0.6269 0.5277 0.7264
No log 2.9149 274 0.5175 0.5816 0.5175 0.7193
No log 2.9362 276 0.4844 0.5675 0.4844 0.6960
No log 2.9574 278 0.4751 0.5589 0.4751 0.6893
No log 2.9787 280 0.5711 0.5066 0.5711 0.7557
No log 3.0 282 0.7883 0.4395 0.7883 0.8879
No log 3.0213 284 0.9014 0.3855 0.9014 0.9494
No log 3.0426 286 0.8319 0.4236 0.8319 0.9121
No log 3.0638 288 0.6560 0.5090 0.6560 0.8099
No log 3.0851 290 0.4909 0.6146 0.4909 0.7007
No log 3.1064 292 0.5062 0.6165 0.5062 0.7115
No log 3.1277 294 0.5084 0.5958 0.5084 0.7130
No log 3.1489 296 0.4776 0.6095 0.4776 0.6911
No log 3.1702 298 0.5020 0.6353 0.5020 0.7085
No log 3.1915 300 0.5850 0.6236 0.5850 0.7648
No log 3.2128 302 0.6298 0.6038 0.6298 0.7936
No log 3.2340 304 0.5905 0.6390 0.5905 0.7685
No log 3.2553 306 0.5438 0.6533 0.5438 0.7374
No log 3.2766 308 0.5077 0.6516 0.5077 0.7126
No log 3.2979 310 0.4730 0.6690 0.4730 0.6877
No log 3.3191 312 0.4638 0.5316 0.4638 0.6810
No log 3.3404 314 0.4792 0.5076 0.4792 0.6922
No log 3.3617 316 0.5127 0.5021 0.5127 0.7160
No log 3.3830 318 0.5382 0.5102 0.5382 0.7336
No log 3.4043 320 0.5127 0.5251 0.5127 0.7160
No log 3.4255 322 0.4869 0.6058 0.4869 0.6978
No log 3.4468 324 0.4948 0.6285 0.4948 0.7034
No log 3.4681 326 0.5583 0.5639 0.5583 0.7472
No log 3.4894 328 0.5722 0.5635 0.5722 0.7564
No log 3.5106 330 0.5521 0.5975 0.5521 0.7430
No log 3.5319 332 0.4959 0.6745 0.4959 0.7042
No log 3.5532 334 0.5140 0.6619 0.5140 0.7169
No log 3.5745 336 0.5124 0.6759 0.5124 0.7158
No log 3.5957 338 0.4769 0.6600 0.4769 0.6906
No log 3.6170 340 0.5188 0.5624 0.5188 0.7203
No log 3.6383 342 0.5470 0.5151 0.5470 0.7396
No log 3.6596 344 0.5334 0.5216 0.5334 0.7303
No log 3.6809 346 0.5004 0.5387 0.5004 0.7074
No log 3.7021 348 0.4929 0.5934 0.4929 0.7021
No log 3.7234 350 0.4901 0.6447 0.4901 0.7001
No log 3.7447 352 0.5012 0.6573 0.5012 0.7079
No log 3.7660 354 0.5848 0.6027 0.5848 0.7647
No log 3.7872 356 0.6120 0.6151 0.6120 0.7823
No log 3.8085 358 0.5581 0.6489 0.5581 0.7471
No log 3.8298 360 0.5284 0.6600 0.5284 0.7269
No log 3.8511 362 0.5164 0.6891 0.5164 0.7186
No log 3.8723 364 0.4974 0.6760 0.4974 0.7052
No log 3.8936 366 0.4727 0.6163 0.4727 0.6875
No log 3.9149 368 0.4944 0.6081 0.4944 0.7032
No log 3.9362 370 0.5499 0.5495 0.5499 0.7415
No log 3.9574 372 0.5458 0.5598 0.5458 0.7388
No log 3.9787 374 0.5039 0.5909 0.5039 0.7099
No log 4.0 376 0.4717 0.5975 0.4717 0.6868
No log 4.0213 378 0.4911 0.5870 0.4911 0.7008
No log 4.0426 380 0.4834 0.5792 0.4834 0.6952
No log 4.0638 382 0.4720 0.5888 0.4720 0.6870
No log 4.0851 384 0.4627 0.5849 0.4627 0.6802
No log 4.1064 386 0.4409 0.6145 0.4409 0.6640
No log 4.1277 388 0.4467 0.6675 0.4467 0.6684
No log 4.1489 390 0.4880 0.6400 0.4880 0.6986
No log 4.1702 392 0.5031 0.6345 0.5031 0.7093
No log 4.1915 394 0.4769 0.6426 0.4769 0.6906
No log 4.2128 396 0.4893 0.6438 0.4893 0.6995
No log 4.2340 398 0.4917 0.6533 0.4917 0.7012
No log 4.2553 400 0.5095 0.6478 0.5095 0.7138
No log 4.2766 402 0.5181 0.6523 0.5181 0.7198
No log 4.2979 404 0.5847 0.6375 0.5847 0.7647
No log 4.3191 406 0.5826 0.6294 0.5826 0.7633
No log 4.3404 408 0.7326 0.5656 0.7326 0.8559
No log 4.3617 410 0.8572 0.4548 0.8572 0.9258
No log 4.3830 412 0.8105 0.3587 0.8105 0.9003
No log 4.4043 414 0.6494 0.4740 0.6494 0.8058
No log 4.4255 416 0.5308 0.5718 0.5308 0.7286
No log 4.4468 418 0.4774 0.6283 0.4774 0.6909
No log 4.4681 420 0.4908 0.6771 0.4908 0.7006
No log 4.4894 422 0.5297 0.6697 0.5297 0.7278
No log 4.5106 424 0.6315 0.5864 0.6315 0.7946
No log 4.5319 426 0.6320 0.5785 0.6320 0.7950
No log 4.5532 428 0.5297 0.5976 0.5297 0.7278
No log 4.5745 430 0.5120 0.5948 0.5120 0.7155
No log 4.5957 432 0.5498 0.5354 0.5498 0.7415
No log 4.6170 434 0.5038 0.5865 0.5038 0.7098
No log 4.6383 436 0.5117 0.5699 0.5117 0.7154
No log 4.6596 438 0.5730 0.5601 0.5730 0.7570
No log 4.6809 440 0.7110 0.4959 0.7110 0.8432
No log 4.7021 442 0.7599 0.4958 0.7599 0.8717
No log 4.7234 444 0.6497 0.5691 0.6497 0.8060
No log 4.7447 446 0.5367 0.6670 0.5367 0.7326
No log 4.7660 448 0.5467 0.6287 0.5467 0.7394
No log 4.7872 450 0.5553 0.6487 0.5553 0.7452
No log 4.8085 452 0.5934 0.6336 0.5934 0.7703
No log 4.8298 454 0.5889 0.6263 0.5889 0.7674
No log 4.8511 456 0.5561 0.6435 0.5561 0.7457
No log 4.8723 458 0.5125 0.6396 0.5125 0.7159
No log 4.8936 460 0.5313 0.6556 0.5313 0.7289
No log 4.9149 462 0.5516 0.6569 0.5516 0.7427
No log 4.9362 464 0.6013 0.6371 0.6013 0.7754
No log 4.9574 466 0.6565 0.6248 0.6565 0.8103
No log 4.9787 468 0.6041 0.6537 0.6041 0.7772
No log 5.0 470 0.5576 0.6111 0.5576 0.7467
No log 5.0213 472 0.5991 0.6003 0.5991 0.7740
No log 5.0426 474 0.6147 0.5890 0.6147 0.7840
No log 5.0638 476 0.6603 0.5319 0.6603 0.8126
No log 5.0851 478 0.6793 0.5204 0.6793 0.8242
No log 5.1064 480 0.5865 0.5492 0.5865 0.7658
No log 5.1277 482 0.5210 0.6362 0.5210 0.7218
No log 5.1489 484 0.5105 0.6234 0.5105 0.7145
No log 5.1702 486 0.5060 0.6503 0.5060 0.7113
No log 5.1915 488 0.5086 0.6532 0.5086 0.7132
No log 5.2128 490 0.5248 0.6547 0.5248 0.7244
No log 5.2340 492 0.5601 0.5923 0.5601 0.7484
No log 5.2553 494 0.5468 0.6075 0.5468 0.7395
No log 5.2766 496 0.4782 0.6802 0.4782 0.6915
No log 5.2979 498 0.4698 0.6618 0.4698 0.6854
0.4903 5.3191 500 0.4830 0.6556 0.4830 0.6950
0.4903 5.3404 502 0.5126 0.6835 0.5126 0.7160
0.4903 5.3617 504 0.5572 0.6616 0.5572 0.7465
0.4903 5.3830 506 0.5345 0.6754 0.5345 0.7311
0.4903 5.4043 508 0.5361 0.5817 0.5361 0.7322
0.4903 5.4255 510 0.5786 0.5911 0.5786 0.7607
0.4903 5.4468 512 0.5646 0.5950 0.5646 0.7514
0.4903 5.4681 514 0.5470 0.6167 0.5470 0.7396
0.4903 5.4894 516 0.4856 0.6977 0.4856 0.6969
0.4903 5.5106 518 0.4482 0.6599 0.4482 0.6695
0.4903 5.5319 520 0.4976 0.6097 0.4976 0.7054
0.4903 5.5532 522 0.4847 0.6042 0.4847 0.6962
0.4903 5.5745 524 0.4543 0.6768 0.4543 0.6740
0.4903 5.5957 526 0.5200 0.6747 0.5200 0.7211
0.4903 5.6170 528 0.5514 0.6245 0.5514 0.7425
0.4903 5.6383 530 0.5425 0.6403 0.5425 0.7366
0.4903 5.6596 532 0.5133 0.6512 0.5133 0.7165
0.4903 5.6809 534 0.5499 0.6483 0.5499 0.7416
0.4903 5.7021 536 0.6707 0.5877 0.6707 0.8190

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask1_vocabulary

Finetuned
(4019)
this model