Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask8_vocabulary

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4874
  • Qwk: 0.7083
  • Mse: 0.4874
  • Rmse: 0.6982

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0196 2 5.0899 0.0138 5.0899 2.2561
No log 0.0392 4 4.2191 0.0315 4.2191 2.0540
No log 0.0588 6 2.5866 0.1001 2.5866 1.6083
No log 0.0784 8 1.1030 0.1550 1.1030 1.0503
No log 0.0980 10 0.9100 0.1859 0.9100 0.9540
No log 0.1176 12 0.8492 0.0878 0.8492 0.9215
No log 0.1373 14 0.8001 0.1409 0.8001 0.8945
No log 0.1569 16 0.7121 0.3458 0.7121 0.8439
No log 0.1765 18 0.6538 0.3626 0.6538 0.8086
No log 0.1961 20 0.6621 0.3121 0.6621 0.8137
No log 0.2157 22 0.7705 0.1907 0.7705 0.8778
No log 0.2353 24 1.0085 0.0504 1.0085 1.0042
No log 0.2549 26 0.9710 0.0740 0.9710 0.9854
No log 0.2745 28 0.7358 0.3412 0.7358 0.8578
No log 0.2941 30 0.6021 0.4423 0.6021 0.7760
No log 0.3137 32 0.6185 0.4610 0.6185 0.7864
No log 0.3333 34 0.6002 0.5080 0.6002 0.7748
No log 0.3529 36 0.6144 0.5144 0.6144 0.7839
No log 0.3725 38 0.6935 0.5210 0.6935 0.8328
No log 0.3922 40 0.6597 0.5757 0.6597 0.8122
No log 0.4118 42 0.5880 0.5986 0.5880 0.7668
No log 0.4314 44 0.5824 0.5943 0.5824 0.7631
No log 0.4510 46 0.5940 0.5805 0.5940 0.7707
No log 0.4706 48 0.5809 0.6174 0.5809 0.7621
No log 0.4902 50 0.7588 0.5854 0.7588 0.8711
No log 0.5098 52 0.7194 0.5921 0.7194 0.8481
No log 0.5294 54 0.6073 0.5430 0.6073 0.7793
No log 0.5490 56 0.6060 0.5059 0.6060 0.7785
No log 0.5686 58 0.6003 0.5017 0.6003 0.7748
No log 0.5882 60 0.6740 0.5066 0.6740 0.8210
No log 0.6078 62 0.6331 0.5265 0.6331 0.7957
No log 0.6275 64 0.6214 0.5549 0.6214 0.7883
No log 0.6471 66 0.5160 0.5756 0.5160 0.7183
No log 0.6667 68 0.4849 0.5951 0.4849 0.6964
No log 0.6863 70 0.4698 0.6265 0.4698 0.6854
No log 0.7059 72 0.5219 0.6463 0.5219 0.7225
No log 0.7255 74 0.5732 0.6315 0.5732 0.7571
No log 0.7451 76 0.6494 0.6067 0.6494 0.8059
No log 0.7647 78 0.6252 0.5981 0.6252 0.7907
No log 0.7843 80 0.5176 0.5970 0.5176 0.7195
No log 0.8039 82 0.5388 0.5712 0.5388 0.7341
No log 0.8235 84 0.5281 0.5969 0.5281 0.7267
No log 0.8431 86 0.5227 0.6105 0.5227 0.7229
No log 0.8627 88 0.7157 0.5914 0.7157 0.8460
No log 0.8824 90 0.7221 0.5927 0.7221 0.8498
No log 0.9020 92 0.5551 0.6063 0.5551 0.7450
No log 0.9216 94 0.4543 0.6202 0.4543 0.6740
No log 0.9412 96 0.4553 0.5847 0.4553 0.6747
No log 0.9608 98 0.4841 0.5466 0.4841 0.6958
No log 0.9804 100 0.5303 0.4554 0.5303 0.7282
No log 1.0 102 0.5424 0.4420 0.5424 0.7365
No log 1.0196 104 0.4888 0.5724 0.4888 0.6991
No log 1.0392 106 0.4636 0.6428 0.4636 0.6809
No log 1.0588 108 0.5058 0.6437 0.5058 0.7112
No log 1.0784 110 0.5212 0.6708 0.5212 0.7219
No log 1.0980 112 0.5112 0.6653 0.5112 0.7150
No log 1.1176 114 0.4826 0.6643 0.4826 0.6947
No log 1.1373 116 0.5358 0.6393 0.5358 0.7320
No log 1.1569 118 0.5526 0.6575 0.5526 0.7434
No log 1.1765 120 0.5272 0.6780 0.5272 0.7261
No log 1.1961 122 0.5880 0.6682 0.5880 0.7668
No log 1.2157 124 0.5921 0.6745 0.5921 0.7695
No log 1.2353 126 0.5485 0.6761 0.5485 0.7406
No log 1.2549 128 0.6278 0.6457 0.6278 0.7923
No log 1.2745 130 0.6787 0.6040 0.6787 0.8238
No log 1.2941 132 0.5832 0.6052 0.5832 0.7636
No log 1.3137 134 0.5542 0.6088 0.5542 0.7444
No log 1.3333 136 0.6077 0.6161 0.6077 0.7796
No log 1.3529 138 0.6055 0.6464 0.6055 0.7782
No log 1.3725 140 0.4621 0.6699 0.4621 0.6798
No log 1.3922 142 0.4722 0.6891 0.4722 0.6871
No log 1.4118 144 0.4934 0.6781 0.4934 0.7024
No log 1.4314 146 0.4966 0.7058 0.4966 0.7047
No log 1.4510 148 0.6699 0.6753 0.6699 0.8185
No log 1.4706 150 0.6961 0.6397 0.6961 0.8343
No log 1.4902 152 0.4944 0.6489 0.4944 0.7032
No log 1.5098 154 0.4541 0.6585 0.4541 0.6739
No log 1.5294 156 0.5172 0.6097 0.5172 0.7192
No log 1.5490 158 0.6770 0.5721 0.6770 0.8228
No log 1.5686 160 0.7086 0.5793 0.7086 0.8418
No log 1.5882 162 0.6065 0.5622 0.6065 0.7788
No log 1.6078 164 0.4984 0.5677 0.4984 0.7060
No log 1.6275 166 0.4309 0.6166 0.4309 0.6564
No log 1.6471 168 0.4174 0.6200 0.4174 0.6461
No log 1.6667 170 0.4667 0.6549 0.4667 0.6831
No log 1.6863 172 0.6216 0.6096 0.6216 0.7884
No log 1.7059 174 0.6465 0.6128 0.6465 0.8040
No log 1.7255 176 0.4906 0.6014 0.4906 0.7004
No log 1.7451 178 0.4340 0.6452 0.4340 0.6588
No log 1.7647 180 0.5022 0.6493 0.5022 0.7087
No log 1.7843 182 0.5293 0.6321 0.5293 0.7275
No log 1.8039 184 0.4863 0.6370 0.4863 0.6974
No log 1.8235 186 0.4378 0.5913 0.4378 0.6617
No log 1.8431 188 0.4734 0.6055 0.4734 0.6880
No log 1.8627 190 0.6173 0.6284 0.6173 0.7857
No log 1.8824 192 0.7387 0.5652 0.7387 0.8595
No log 1.9020 194 0.5997 0.6477 0.5997 0.7744
No log 1.9216 196 0.4444 0.7054 0.4444 0.6666
No log 1.9412 198 0.4395 0.6947 0.4395 0.6629
No log 1.9608 200 0.4316 0.6947 0.4316 0.6569
No log 1.9804 202 0.4581 0.6840 0.4581 0.6768
No log 2.0 204 0.4475 0.6491 0.4475 0.6690
No log 2.0196 206 0.4144 0.6651 0.4144 0.6437
No log 2.0392 208 0.4096 0.6950 0.4096 0.6400
No log 2.0588 210 0.4350 0.7113 0.4350 0.6595
No log 2.0784 212 0.5867 0.6692 0.5867 0.7660
No log 2.0980 214 0.6919 0.6160 0.6919 0.8318
No log 2.1176 216 0.6905 0.5966 0.6905 0.8309
No log 2.1373 218 0.5215 0.6485 0.5215 0.7222
No log 2.1569 220 0.4969 0.6382 0.4969 0.7049
No log 2.1765 222 0.4945 0.6304 0.4945 0.7032
No log 2.1961 224 0.4516 0.6495 0.4516 0.6720
No log 2.2157 226 0.4397 0.6602 0.4397 0.6631
No log 2.2353 228 0.4506 0.6531 0.4506 0.6713
No log 2.2549 230 0.5033 0.6547 0.5033 0.7094
No log 2.2745 232 0.4893 0.6606 0.4893 0.6995
No log 2.2941 234 0.5293 0.6752 0.5293 0.7275
No log 2.3137 236 0.6836 0.5587 0.6836 0.8268
No log 2.3333 238 0.7270 0.5659 0.7270 0.8526
No log 2.3529 240 0.6155 0.6457 0.6155 0.7845
No log 2.3725 242 0.4908 0.6913 0.4908 0.7005
No log 2.3922 244 0.5442 0.6703 0.5442 0.7377
No log 2.4118 246 0.5839 0.7018 0.5839 0.7641
No log 2.4314 248 0.8137 0.6769 0.8137 0.9021
No log 2.4510 250 0.7385 0.6900 0.7385 0.8594
No log 2.4706 252 0.5235 0.7221 0.5235 0.7235
No log 2.4902 254 0.4364 0.6669 0.4364 0.6606
No log 2.5098 256 0.4331 0.6235 0.4331 0.6581
No log 2.5294 258 0.5979 0.5910 0.5979 0.7733
No log 2.5490 260 0.9717 0.4166 0.9717 0.9857
No log 2.5686 262 1.0300 0.4020 1.0300 1.0149
No log 2.5882 264 0.8110 0.4899 0.8110 0.9005
No log 2.6078 266 0.5805 0.5708 0.5805 0.7619
No log 2.6275 268 0.5102 0.5752 0.5102 0.7143
No log 2.6471 270 0.4466 0.6256 0.4466 0.6683
No log 2.6667 272 0.4379 0.6457 0.4379 0.6617
No log 2.6863 274 0.4670 0.6698 0.4670 0.6834
No log 2.7059 276 0.5285 0.6570 0.5285 0.7270
No log 2.7255 278 0.5080 0.6568 0.5080 0.7128
No log 2.7451 280 0.4701 0.7071 0.4701 0.6856
No log 2.7647 282 0.4805 0.6471 0.4805 0.6932
No log 2.7843 284 0.4891 0.6543 0.4891 0.6994
No log 2.8039 286 0.4752 0.6821 0.4752 0.6893
No log 2.8235 288 0.4947 0.6643 0.4947 0.7034
No log 2.8431 290 0.6934 0.6377 0.6934 0.8327
No log 2.8627 292 0.8227 0.5569 0.8227 0.9070
No log 2.8824 294 0.7619 0.5712 0.7619 0.8729
No log 2.9020 296 0.5097 0.6135 0.5097 0.7139
No log 2.9216 298 0.4599 0.6074 0.4599 0.6782
No log 2.9412 300 0.4637 0.5726 0.4637 0.6809
No log 2.9608 302 0.4807 0.5656 0.4807 0.6933
No log 2.9804 304 0.5346 0.5521 0.5346 0.7312
No log 3.0 306 0.7054 0.5395 0.7054 0.8399
No log 3.0196 308 0.7197 0.5687 0.7197 0.8483
No log 3.0392 310 0.5696 0.6057 0.5696 0.7547
No log 3.0588 312 0.5280 0.6268 0.5280 0.7267
No log 3.0784 314 0.5879 0.6221 0.5879 0.7667
No log 3.0980 316 0.5059 0.6743 0.5059 0.7113
No log 3.1176 318 0.5160 0.6841 0.5160 0.7183
No log 3.1373 320 0.6853 0.6311 0.6853 0.8278
No log 3.1569 322 0.8771 0.5688 0.8771 0.9365
No log 3.1765 324 0.7364 0.5798 0.7364 0.8582
No log 3.1961 326 0.5266 0.6430 0.5266 0.7256
No log 3.2157 328 0.4472 0.6986 0.4472 0.6687
No log 3.2353 330 0.4470 0.6899 0.4470 0.6686
No log 3.2549 332 0.4667 0.6628 0.4667 0.6831
No log 3.2745 334 0.6336 0.5992 0.6336 0.7960
No log 3.2941 336 0.6680 0.5900 0.6680 0.8173
No log 3.3137 338 0.7144 0.5631 0.7144 0.8452
No log 3.3333 340 0.6633 0.6150 0.6633 0.8144
No log 3.3529 342 0.5403 0.6620 0.5403 0.7350
No log 3.3725 344 0.4881 0.6869 0.4881 0.6986
No log 3.3922 346 0.5613 0.6681 0.5613 0.7492
No log 3.4118 348 0.8000 0.6113 0.8000 0.8944
No log 3.4314 350 1.1641 0.5456 1.1641 1.0789
No log 3.4510 352 0.9233 0.6014 0.9233 0.9609
No log 3.4706 354 0.4484 0.6573 0.4484 0.6696
No log 3.4902 356 0.4162 0.6552 0.4162 0.6451
No log 3.5098 358 0.5994 0.6294 0.5994 0.7742
No log 3.5294 360 0.9521 0.5993 0.9521 0.9758
No log 3.5490 362 0.7894 0.6178 0.7894 0.8885
No log 3.5686 364 0.4678 0.6853 0.4678 0.6839
No log 3.5882 366 0.4239 0.7196 0.4239 0.6510
No log 3.6078 368 0.4547 0.7172 0.4547 0.6743
No log 3.6275 370 0.7712 0.6668 0.7712 0.8782
No log 3.6471 372 1.3317 0.4993 1.3317 1.1540
No log 3.6667 374 1.3701 0.4824 1.3701 1.1705
No log 3.6863 376 1.0346 0.5881 1.0346 1.0171
No log 3.7059 378 0.6191 0.6995 0.6191 0.7868
No log 3.7255 380 0.4948 0.7101 0.4948 0.7034
No log 3.7451 382 0.4476 0.7245 0.4476 0.6690
No log 3.7647 384 0.4349 0.6980 0.4349 0.6595
No log 3.7843 386 0.5607 0.6153 0.5607 0.7488
No log 3.8039 388 0.7080 0.5753 0.7080 0.8414
No log 3.8235 390 0.7205 0.5582 0.7205 0.8488
No log 3.8431 392 0.6220 0.5785 0.6220 0.7886
No log 3.8627 394 0.5408 0.6275 0.5408 0.7354
No log 3.8824 396 0.5583 0.6698 0.5583 0.7472
No log 3.9020 398 0.5236 0.7094 0.5236 0.7236
No log 3.9216 400 0.5440 0.6910 0.5440 0.7376
No log 3.9412 402 0.5721 0.6994 0.5721 0.7564
No log 3.9608 404 0.7204 0.7026 0.7204 0.8488
No log 3.9804 406 0.6952 0.6917 0.6952 0.8338
No log 4.0 408 0.5650 0.7142 0.5650 0.7516
No log 4.0196 410 0.5183 0.7078 0.5183 0.7199
No log 4.0392 412 0.5525 0.6946 0.5525 0.7433
No log 4.0588 414 0.5282 0.6945 0.5282 0.7268
No log 4.0784 416 0.4845 0.6844 0.4845 0.6960
No log 4.0980 418 0.4362 0.7045 0.4362 0.6605
No log 4.1176 420 0.4690 0.6600 0.4690 0.6848
No log 4.1373 422 0.6592 0.6057 0.6592 0.8119
No log 4.1569 424 0.9423 0.5024 0.9423 0.9707
No log 4.1765 426 0.8644 0.5283 0.8644 0.9297
No log 4.1961 428 0.6554 0.6028 0.6554 0.8095
No log 4.2157 430 0.4676 0.6599 0.4676 0.6838
No log 4.2353 432 0.4300 0.6753 0.4300 0.6557
No log 4.2549 434 0.4513 0.6534 0.4513 0.6718
No log 4.2745 436 0.5305 0.6299 0.5305 0.7284
No log 4.2941 438 0.4958 0.6315 0.4958 0.7041
No log 4.3137 440 0.4720 0.6247 0.4720 0.6870
No log 4.3333 442 0.5119 0.6443 0.5119 0.7155
No log 4.3529 444 0.5003 0.6746 0.5003 0.7073
No log 4.3725 446 0.4285 0.7154 0.4285 0.6546
No log 4.3922 448 0.4340 0.7406 0.4340 0.6588
No log 4.4118 450 0.5131 0.7364 0.5131 0.7163
No log 4.4314 452 0.5764 0.7038 0.5764 0.7592
No log 4.4510 454 0.6457 0.6833 0.6457 0.8036
No log 4.4706 456 0.7289 0.6457 0.7289 0.8538
No log 4.4902 458 0.6970 0.6625 0.6970 0.8349
No log 4.5098 460 0.4602 0.7153 0.4602 0.6784
No log 4.5294 462 0.4440 0.6954 0.4440 0.6663
No log 4.5490 464 0.5276 0.6647 0.5276 0.7263
No log 4.5686 466 0.8252 0.5565 0.8252 0.9084
No log 4.5882 468 0.9114 0.5407 0.9114 0.9547
No log 4.6078 470 0.6325 0.6478 0.6325 0.7953
No log 4.6275 472 0.5418 0.6764 0.5418 0.7361
No log 4.6471 474 0.5526 0.6671 0.5526 0.7434
No log 4.6667 476 0.5176 0.6562 0.5176 0.7194
No log 4.6863 478 0.4760 0.6594 0.4760 0.6899
No log 4.7059 480 0.5836 0.6331 0.5836 0.7639
No log 4.7255 482 0.6288 0.6202 0.6288 0.7930
No log 4.7451 484 0.5975 0.6346 0.5975 0.7730
No log 4.7647 486 0.4696 0.6430 0.4696 0.6853
No log 4.7843 488 0.4513 0.6484 0.4513 0.6718
No log 4.8039 490 0.4812 0.6763 0.4812 0.6937
No log 4.8235 492 0.6105 0.6315 0.6105 0.7813
No log 4.8431 494 0.7897 0.5842 0.7897 0.8887
No log 4.8627 496 0.6601 0.6432 0.6601 0.8125
No log 4.8824 498 0.5147 0.6864 0.5147 0.7174
0.5079 4.9020 500 0.4278 0.7398 0.4278 0.6541
0.5079 4.9216 502 0.4733 0.7230 0.4733 0.6880
0.5079 4.9412 504 0.5510 0.6638 0.5510 0.7423
0.5079 4.9608 506 0.7516 0.5977 0.7516 0.8670
0.5079 4.9804 508 0.7260 0.6083 0.7260 0.8521
0.5079 5.0 510 0.4874 0.7083 0.4874 0.6982

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask8_vocabulary

Finetuned
(4019)
this model