Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask5_vocabulary

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4963
  • Qwk: 0.5684
  • Mse: 0.4963
  • Rmse: 0.7045

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0198 2 4.4542 -0.0037 4.4542 2.1105
No log 0.0396 4 3.3121 0.0462 3.3121 1.8199
No log 0.0594 6 1.6638 -0.0003 1.6638 1.2899
No log 0.0792 8 0.9917 0.1036 0.9917 0.9958
No log 0.0990 10 0.7569 0.0900 0.7569 0.8700
No log 0.1188 12 0.8518 0.0291 0.8518 0.9230
No log 0.1386 14 0.9221 -0.0131 0.9221 0.9603
No log 0.1584 16 0.9190 -0.0155 0.9190 0.9587
No log 0.1782 18 0.8911 -0.0755 0.8911 0.9440
No log 0.1980 20 0.8198 0.0249 0.8198 0.9054
No log 0.2178 22 0.7043 0.2336 0.7043 0.8392
No log 0.2376 24 0.5999 0.3270 0.5999 0.7745
No log 0.2574 26 0.6430 0.4107 0.6430 0.8019
No log 0.2772 28 0.6274 0.3699 0.6274 0.7921
No log 0.2970 30 0.6022 0.3127 0.6022 0.7760
No log 0.3168 32 0.6329 0.2912 0.6329 0.7956
No log 0.3366 34 0.6689 0.2554 0.6689 0.8179
No log 0.3564 36 0.6672 0.4415 0.6672 0.8168
No log 0.3762 38 0.6585 0.4561 0.6585 0.8115
No log 0.3960 40 0.6783 0.5408 0.6783 0.8236
No log 0.4158 42 0.6391 0.5998 0.6391 0.7995
No log 0.4356 44 0.5826 0.5048 0.5826 0.7633
No log 0.4554 46 0.5991 0.3990 0.5991 0.7740
No log 0.4752 48 0.6826 0.2625 0.6826 0.8262
No log 0.4950 50 0.6797 0.3275 0.6797 0.8245
No log 0.5149 52 0.5910 0.4762 0.5910 0.7687
No log 0.5347 54 0.4986 0.5375 0.4986 0.7061
No log 0.5545 56 0.4903 0.5172 0.4903 0.7002
No log 0.5743 58 0.5548 0.5065 0.5548 0.7448
No log 0.5941 60 0.6378 0.4624 0.6378 0.7986
No log 0.6139 62 0.6827 0.5108 0.6827 0.8263
No log 0.6337 64 0.6637 0.5864 0.6637 0.8147
No log 0.6535 66 0.7336 0.6156 0.7336 0.8565
No log 0.6733 68 0.7254 0.5908 0.7254 0.8517
No log 0.6931 70 0.6467 0.5684 0.6467 0.8042
No log 0.7129 72 0.5111 0.6422 0.5111 0.7149
No log 0.7327 74 0.4766 0.6189 0.4766 0.6904
No log 0.7525 76 0.4270 0.5915 0.4270 0.6535
No log 0.7723 78 0.4342 0.4992 0.4342 0.6589
No log 0.7921 80 0.4572 0.4361 0.4572 0.6762
No log 0.8119 82 0.5485 0.4110 0.5485 0.7406
No log 0.8317 84 0.6427 0.2669 0.6427 0.8017
No log 0.8515 86 0.5813 0.3211 0.5813 0.7624
No log 0.8713 88 0.5300 0.3696 0.5300 0.7280
No log 0.8911 90 0.4984 0.4030 0.4984 0.7060
No log 0.9109 92 0.4602 0.4381 0.4602 0.6783
No log 0.9307 94 0.4293 0.5140 0.4293 0.6552
No log 0.9505 96 0.4625 0.5609 0.4625 0.6801
No log 0.9703 98 0.5080 0.5713 0.5080 0.7128
No log 0.9901 100 0.5215 0.5982 0.5215 0.7221
No log 1.0099 102 0.5515 0.5895 0.5515 0.7426
No log 1.0297 104 0.5372 0.5690 0.5372 0.7330
No log 1.0495 106 0.5098 0.5581 0.5098 0.7140
No log 1.0693 108 0.4765 0.6006 0.4765 0.6903
No log 1.0891 110 0.5419 0.5427 0.5419 0.7361
No log 1.1089 112 0.5417 0.5364 0.5417 0.7360
No log 1.1287 114 0.5281 0.5477 0.5281 0.7267
No log 1.1485 116 0.5352 0.5434 0.5352 0.7316
No log 1.1683 118 0.4610 0.5633 0.4610 0.6790
No log 1.1881 120 0.4257 0.5686 0.4257 0.6525
No log 1.2079 122 0.4849 0.5469 0.4849 0.6964
No log 1.2277 124 0.5815 0.5354 0.5815 0.7626
No log 1.2475 126 0.5383 0.5602 0.5383 0.7337
No log 1.2673 128 0.4626 0.6708 0.4626 0.6802
No log 1.2871 130 0.5182 0.6520 0.5182 0.7198
No log 1.3069 132 0.5259 0.6537 0.5259 0.7252
No log 1.3267 134 0.7994 0.5207 0.7994 0.8941
No log 1.3465 136 0.9838 0.3583 0.9838 0.9919
No log 1.3663 138 0.8917 0.3556 0.8917 0.9443
No log 1.3861 140 0.6958 0.3589 0.6958 0.8342
No log 1.4059 142 0.5785 0.3673 0.5785 0.7606
No log 1.4257 144 0.5319 0.3924 0.5319 0.7293
No log 1.4455 146 0.4598 0.4717 0.4598 0.6781
No log 1.4653 148 0.4437 0.5503 0.4437 0.6661
No log 1.4851 150 0.4373 0.5778 0.4373 0.6613
No log 1.5050 152 0.5175 0.5694 0.5175 0.7194
No log 1.5248 154 0.6907 0.5298 0.6907 0.8311
No log 1.5446 156 0.8126 0.4469 0.8126 0.9015
No log 1.5644 158 0.7276 0.4594 0.7276 0.8530
No log 1.5842 160 0.6637 0.4829 0.6637 0.8147
No log 1.6040 162 0.6331 0.5213 0.6331 0.7957
No log 1.6238 164 0.5688 0.5613 0.5688 0.7542
No log 1.6436 166 0.5327 0.5889 0.5327 0.7298
No log 1.6634 168 0.5711 0.5717 0.5711 0.7557
No log 1.6832 170 0.6696 0.4869 0.6696 0.8183
No log 1.7030 172 0.6125 0.5151 0.6125 0.7826
No log 1.7228 174 0.4620 0.5595 0.4620 0.6797
No log 1.7426 176 0.4127 0.5848 0.4127 0.6424
No log 1.7624 178 0.4144 0.6294 0.4144 0.6437
No log 1.7822 180 0.4630 0.5959 0.4630 0.6804
No log 1.8020 182 0.5589 0.5531 0.5589 0.7476
No log 1.8218 184 0.5760 0.5470 0.5760 0.7589
No log 1.8416 186 0.5757 0.5554 0.5757 0.7587
No log 1.8614 188 0.5340 0.6139 0.5340 0.7307
No log 1.8812 190 0.4959 0.6266 0.4959 0.7042
No log 1.9010 192 0.4857 0.5955 0.4857 0.6969
No log 1.9208 194 0.5166 0.6129 0.5166 0.7188
No log 1.9406 196 0.5253 0.5982 0.5253 0.7248
No log 1.9604 198 0.5117 0.6336 0.5117 0.7153
No log 1.9802 200 0.5096 0.6542 0.5096 0.7139
No log 2.0 202 0.6029 0.6154 0.6029 0.7765
No log 2.0198 204 0.6022 0.6162 0.6022 0.7760
No log 2.0396 206 0.5900 0.6120 0.5900 0.7681
No log 2.0594 208 0.5069 0.6468 0.5069 0.7120
No log 2.0792 210 0.4931 0.6359 0.4931 0.7022
No log 2.0990 212 0.5573 0.5886 0.5573 0.7465
No log 2.1188 214 0.6994 0.5214 0.6994 0.8363
No log 2.1386 216 0.6707 0.5101 0.6707 0.8190
No log 2.1584 218 0.5328 0.5484 0.5328 0.7299
No log 2.1782 220 0.4572 0.5784 0.4572 0.6762
No log 2.1980 222 0.4924 0.5432 0.4924 0.7017
No log 2.2178 224 0.6986 0.4858 0.6986 0.8358
No log 2.2376 226 0.8505 0.4735 0.8505 0.9222
No log 2.2574 228 0.7855 0.4896 0.7855 0.8863
No log 2.2772 230 0.7117 0.5509 0.7117 0.8436
No log 2.2970 232 0.6060 0.6039 0.6060 0.7785
No log 2.3168 234 0.5017 0.6628 0.5017 0.7083
No log 2.3366 236 0.4612 0.6801 0.4612 0.6791
No log 2.3564 238 0.4147 0.6544 0.4147 0.6440
No log 2.3762 240 0.4110 0.6477 0.4110 0.6411
No log 2.3960 242 0.4261 0.5995 0.4261 0.6527
No log 2.4158 244 0.4637 0.5663 0.4637 0.6810
No log 2.4356 246 0.6081 0.5611 0.6081 0.7798
No log 2.4554 248 0.6584 0.5661 0.6584 0.8114
No log 2.4752 250 0.5716 0.6149 0.5716 0.7560
No log 2.4950 252 0.4598 0.6266 0.4598 0.6781
No log 2.5149 254 0.4262 0.6271 0.4262 0.6528
No log 2.5347 256 0.4143 0.6533 0.4143 0.6436
No log 2.5545 258 0.4333 0.6506 0.4333 0.6583
No log 2.5743 260 0.4537 0.6602 0.4537 0.6736
No log 2.5941 262 0.4570 0.6382 0.4570 0.6760
No log 2.6139 264 0.4060 0.6558 0.4060 0.6372
No log 2.6337 266 0.4006 0.6923 0.4006 0.6329
No log 2.6535 268 0.4163 0.6881 0.4163 0.6452
No log 2.6733 270 0.4278 0.6364 0.4278 0.6541
No log 2.6931 272 0.4108 0.6486 0.4108 0.6409
No log 2.7129 274 0.4050 0.6712 0.4050 0.6364
No log 2.7327 276 0.4292 0.6576 0.4292 0.6551
No log 2.7525 278 0.5614 0.5904 0.5614 0.7493
No log 2.7723 280 0.7061 0.5506 0.7061 0.8403
No log 2.7921 282 0.6979 0.5927 0.6979 0.8354
No log 2.8119 284 0.5830 0.6391 0.5830 0.7636
No log 2.8317 286 0.5023 0.6222 0.5023 0.7088
No log 2.8515 288 0.5225 0.6175 0.5225 0.7229
No log 2.8713 290 0.4655 0.6312 0.4655 0.6823
No log 2.8911 292 0.4762 0.5916 0.4762 0.6901
No log 2.9109 294 0.6171 0.5565 0.6171 0.7856
No log 2.9307 296 0.6314 0.5761 0.6314 0.7946
No log 2.9505 298 0.5483 0.5619 0.5483 0.7405
No log 2.9703 300 0.5081 0.5973 0.5081 0.7128
No log 2.9901 302 0.4677 0.6211 0.4677 0.6839
No log 3.0099 304 0.4252 0.6211 0.4252 0.6521
No log 3.0297 306 0.4183 0.6173 0.4183 0.6467
No log 3.0495 308 0.4219 0.5921 0.4219 0.6496
No log 3.0693 310 0.4559 0.5563 0.4559 0.6752
No log 3.0891 312 0.5116 0.5569 0.5116 0.7152
No log 3.1089 314 0.5252 0.5005 0.5252 0.7247
No log 3.1287 316 0.5468 0.5140 0.5468 0.7394
No log 3.1485 318 0.5175 0.5957 0.5175 0.7194
No log 3.1683 320 0.4538 0.6453 0.4538 0.6736
No log 3.1881 322 0.4331 0.6630 0.4331 0.6581
No log 3.2079 324 0.4694 0.6687 0.4694 0.6851
No log 3.2277 326 0.4827 0.6692 0.4827 0.6948
No log 3.2475 328 0.5059 0.6446 0.5059 0.7112
No log 3.2673 330 0.4903 0.6205 0.4903 0.7002
No log 3.2871 332 0.4567 0.5840 0.4567 0.6758
No log 3.3069 334 0.4878 0.6052 0.4878 0.6984
No log 3.3267 336 0.5594 0.5944 0.5594 0.7479
No log 3.3465 338 0.5464 0.6020 0.5464 0.7392
No log 3.3663 340 0.4364 0.6563 0.4364 0.6606
No log 3.3861 342 0.4167 0.6707 0.4167 0.6455
No log 3.4059 344 0.4716 0.6570 0.4716 0.6867
No log 3.4257 346 0.4945 0.6443 0.4945 0.7032
No log 3.4455 348 0.5325 0.6175 0.5325 0.7297
No log 3.4653 350 0.4789 0.6513 0.4789 0.6921
No log 3.4851 352 0.4735 0.6688 0.4735 0.6881
No log 3.5050 354 0.5005 0.6579 0.5005 0.7074
No log 3.5248 356 0.6506 0.5541 0.6506 0.8066
No log 3.5446 358 0.6375 0.5610 0.6375 0.7984
No log 3.5644 360 0.5978 0.5625 0.5978 0.7732
No log 3.5842 362 0.5105 0.5267 0.5105 0.7145
No log 3.6040 364 0.5826 0.5380 0.5826 0.7633
No log 3.6238 366 0.8995 0.4660 0.8995 0.9484
No log 3.6436 368 0.9632 0.4797 0.9632 0.9814
No log 3.6634 370 0.7670 0.5897 0.7670 0.8758
No log 3.6832 372 0.5770 0.6386 0.5770 0.7596
No log 3.7030 374 0.5678 0.6367 0.5678 0.7535
No log 3.7228 376 0.6025 0.6043 0.6025 0.7762
No log 3.7426 378 0.5414 0.5759 0.5414 0.7358
No log 3.7624 380 0.5327 0.5618 0.5327 0.7299
No log 3.7822 382 0.5559 0.5580 0.5559 0.7456
No log 3.8020 384 0.7227 0.5414 0.7227 0.8501
No log 3.8218 386 1.0514 0.3974 1.0514 1.0254
No log 3.8416 388 1.1877 0.3849 1.1877 1.0898
No log 3.8614 390 0.9528 0.4224 0.9528 0.9761
No log 3.8812 392 0.5446 0.5887 0.5446 0.7380
No log 3.9010 394 0.3866 0.6319 0.3866 0.6218
No log 3.9208 396 0.3848 0.6551 0.3848 0.6203
No log 3.9406 398 0.4565 0.6589 0.4565 0.6757
No log 3.9604 400 0.7295 0.5246 0.7295 0.8541
No log 3.9802 402 0.8713 0.4619 0.8713 0.9334
No log 4.0 404 0.8333 0.4868 0.8333 0.9129
No log 4.0198 406 0.6810 0.5395 0.6810 0.8252
No log 4.0396 408 0.5101 0.6500 0.5101 0.7142
No log 4.0594 410 0.4613 0.6261 0.4613 0.6792
No log 4.0792 412 0.4846 0.6023 0.4846 0.6961
No log 4.0990 414 0.5878 0.5499 0.5878 0.7667
No log 4.1188 416 0.8258 0.4651 0.8258 0.9087
No log 4.1386 418 0.9974 0.3743 0.9974 0.9987
No log 4.1584 420 1.0479 0.3502 1.0479 1.0237
No log 4.1782 422 0.8910 0.4371 0.8910 0.9439
No log 4.1980 424 0.6018 0.5506 0.6018 0.7758
No log 4.2178 426 0.4163 0.6795 0.4163 0.6452
No log 4.2376 428 0.4026 0.6673 0.4026 0.6345
No log 4.2574 430 0.4371 0.6932 0.4371 0.6611
No log 4.2772 432 0.4463 0.6818 0.4463 0.6680
No log 4.2970 434 0.4684 0.6637 0.4684 0.6844
No log 4.3168 436 0.5667 0.6242 0.5667 0.7528
No log 4.3366 438 0.6217 0.5713 0.6217 0.7885
No log 4.3564 440 0.5756 0.5741 0.5756 0.7587
No log 4.3762 442 0.6037 0.5432 0.6037 0.7770
No log 4.3960 444 0.5338 0.5550 0.5338 0.7306
No log 4.4158 446 0.5505 0.5530 0.5505 0.7420
No log 4.4356 448 0.5445 0.5789 0.5445 0.7379
No log 4.4554 450 0.5348 0.5877 0.5348 0.7313
No log 4.4752 452 0.5407 0.5802 0.5407 0.7353
No log 4.4950 454 0.4723 0.6132 0.4723 0.6873
No log 4.5149 456 0.4481 0.6307 0.4481 0.6694
No log 4.5347 458 0.4906 0.6062 0.4906 0.7004
No log 4.5545 460 0.6137 0.5836 0.6137 0.7834
No log 4.5743 462 0.6904 0.5827 0.6904 0.8309
No log 4.5941 464 0.7086 0.5953 0.7086 0.8418
No log 4.6139 466 0.6434 0.6149 0.6434 0.8021
No log 4.6337 468 0.5522 0.6449 0.5522 0.7431
No log 4.6535 470 0.5318 0.6412 0.5318 0.7293
No log 4.6733 472 0.4532 0.6736 0.4532 0.6732
No log 4.6931 474 0.5278 0.6077 0.5278 0.7265
No log 4.7129 476 0.6232 0.5483 0.6232 0.7894
No log 4.7327 478 0.4802 0.5804 0.4802 0.6930
No log 4.7525 480 0.4747 0.5930 0.4747 0.6890
No log 4.7723 482 0.6283 0.5140 0.6283 0.7927
No log 4.7921 484 0.6484 0.5206 0.6484 0.8052
No log 4.8119 486 0.5671 0.5538 0.5671 0.7530
No log 4.8317 488 0.4935 0.6150 0.4935 0.7025
No log 4.8515 490 0.5876 0.5852 0.5876 0.7666
No log 4.8713 492 0.7904 0.5321 0.7904 0.8890
No log 4.8911 494 0.7317 0.5854 0.7317 0.8554
No log 4.9109 496 0.5367 0.6700 0.5367 0.7326
No log 4.9307 498 0.5047 0.6684 0.5047 0.7104
0.5174 4.9505 500 0.4917 0.6803 0.4917 0.7012
0.5174 4.9703 502 0.6110 0.5891 0.6110 0.7816
0.5174 4.9901 504 0.8514 0.5163 0.8514 0.9227
0.5174 5.0099 506 0.8258 0.5016 0.8258 0.9087
0.5174 5.0297 508 0.5840 0.5680 0.5840 0.7642
0.5174 5.0495 510 0.4963 0.5684 0.4963 0.7045

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask5_vocabulary

Finetuned
(4019)
this model