Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask8_development

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4744
  • Qwk: 0.5496
  • Mse: 0.4744
  • Rmse: 0.6887

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0196 2 4.1841 0.0131 4.1841 2.0455
No log 0.0392 4 3.3694 0.0367 3.3694 1.8356
No log 0.0588 6 1.9011 0.1218 1.9011 1.3788
No log 0.0784 8 0.8092 0.1659 0.8092 0.8996
No log 0.0980 10 0.7372 0.1594 0.7372 0.8586
No log 0.1176 12 0.7884 0.0895 0.7884 0.8879
No log 0.1373 14 0.7668 0.3157 0.7668 0.8756
No log 0.1569 16 0.7860 0.2846 0.7860 0.8866
No log 0.1765 18 0.6591 0.2790 0.6591 0.8119
No log 0.1961 20 0.6420 0.2728 0.6420 0.8012
No log 0.2157 22 0.6233 0.3185 0.6233 0.7895
No log 0.2353 24 0.6843 0.1790 0.6843 0.8272
No log 0.2549 26 0.6878 0.2016 0.6878 0.8293
No log 0.2745 28 0.6001 0.3500 0.6001 0.7747
No log 0.2941 30 0.5105 0.4622 0.5105 0.7145
No log 0.3137 32 0.5104 0.5258 0.5104 0.7145
No log 0.3333 34 0.5323 0.5412 0.5323 0.7296
No log 0.3529 36 0.5752 0.5265 0.5752 0.7584
No log 0.3725 38 0.7106 0.4847 0.7106 0.8430
No log 0.3922 40 0.9041 0.2287 0.9041 0.9508
No log 0.4118 42 1.0775 0.0384 1.0775 1.0380
No log 0.4314 44 1.3162 -0.1270 1.3162 1.1473
No log 0.4510 46 1.4042 -0.0351 1.4042 1.1850
No log 0.4706 48 1.1193 0.0766 1.1193 1.0580
No log 0.4902 50 0.7738 0.2692 0.7738 0.8796
No log 0.5098 52 0.6622 0.5501 0.6622 0.8137
No log 0.5294 54 0.5766 0.5594 0.5766 0.7594
No log 0.5490 56 0.5125 0.4951 0.5125 0.7159
No log 0.5686 58 0.6476 0.3506 0.6476 0.8047
No log 0.5882 60 0.8667 0.1810 0.8667 0.9310
No log 0.6078 62 0.9499 0.2013 0.9499 0.9746
No log 0.6275 64 0.9296 0.3390 0.9296 0.9642
No log 0.6471 66 0.8488 0.3626 0.8488 0.9213
No log 0.6667 68 0.6706 0.4117 0.6706 0.8189
No log 0.6863 70 0.5390 0.4752 0.5390 0.7341
No log 0.7059 72 0.5347 0.5345 0.5347 0.7312
No log 0.7255 74 0.5187 0.5375 0.5187 0.7202
No log 0.7451 76 0.5044 0.4763 0.5044 0.7102
No log 0.7647 78 0.5615 0.3887 0.5615 0.7494
No log 0.7843 80 0.5810 0.3887 0.5810 0.7623
No log 0.8039 82 0.5343 0.3959 0.5343 0.7309
No log 0.8235 84 0.5005 0.4515 0.5005 0.7074
No log 0.8431 86 0.4904 0.4729 0.4904 0.7003
No log 0.8627 88 0.4824 0.4530 0.4824 0.6946
No log 0.8824 90 0.4866 0.4353 0.4866 0.6975
No log 0.9020 92 0.5003 0.4346 0.5003 0.7073
No log 0.9216 94 0.4784 0.4640 0.4784 0.6917
No log 0.9412 96 0.4885 0.5290 0.4885 0.6989
No log 0.9608 98 0.4879 0.5156 0.4879 0.6985
No log 0.9804 100 0.4785 0.4936 0.4785 0.6917
No log 1.0 102 0.4815 0.4512 0.4815 0.6939
No log 1.0196 104 0.4857 0.5196 0.4857 0.6969
No log 1.0392 106 0.4887 0.5268 0.4887 0.6990
No log 1.0588 108 0.5089 0.4412 0.5089 0.7134
No log 1.0784 110 0.5336 0.4314 0.5336 0.7305
No log 1.0980 112 0.5377 0.4244 0.5377 0.7333
No log 1.1176 114 0.5137 0.4316 0.5137 0.7168
No log 1.1373 116 0.4879 0.4468 0.4879 0.6985
No log 1.1569 118 0.4711 0.4601 0.4711 0.6863
No log 1.1765 120 0.4621 0.5111 0.4621 0.6798
No log 1.1961 122 0.4667 0.5566 0.4667 0.6831
No log 1.2157 124 0.4836 0.5890 0.4836 0.6954
No log 1.2353 126 0.4990 0.6142 0.4990 0.7064
No log 1.2549 128 0.4753 0.5914 0.4753 0.6894
No log 1.2745 130 0.4727 0.5242 0.4727 0.6876
No log 1.2941 132 0.5021 0.4952 0.5021 0.7086
No log 1.3137 134 0.5289 0.4505 0.5289 0.7272
No log 1.3333 136 0.5100 0.5243 0.5100 0.7141
No log 1.3529 138 0.4521 0.5791 0.4521 0.6724
No log 1.3725 140 0.4599 0.5734 0.4599 0.6782
No log 1.3922 142 0.4724 0.5813 0.4724 0.6873
No log 1.4118 144 0.4366 0.5448 0.4366 0.6608
No log 1.4314 146 0.4612 0.4622 0.4612 0.6791
No log 1.4510 148 0.4764 0.4448 0.4764 0.6902
No log 1.4706 150 0.4935 0.4061 0.4935 0.7025
No log 1.4902 152 0.4972 0.3890 0.4972 0.7051
No log 1.5098 154 0.5366 0.3545 0.5366 0.7325
No log 1.5294 156 0.5402 0.4241 0.5402 0.7350
No log 1.5490 158 0.6181 0.4375 0.6181 0.7862
No log 1.5686 160 0.5969 0.4819 0.5969 0.7726
No log 1.5882 162 0.5734 0.4808 0.5734 0.7572
No log 1.6078 164 0.5033 0.5482 0.5033 0.7094
No log 1.6275 166 0.4178 0.5609 0.4178 0.6464
No log 1.6471 168 0.3978 0.5996 0.3978 0.6307
No log 1.6667 170 0.4323 0.6139 0.4323 0.6575
No log 1.6863 172 0.6541 0.4722 0.6541 0.8088
No log 1.7059 174 0.8573 0.2931 0.8573 0.9259
No log 1.7255 176 0.8885 0.2441 0.8885 0.9426
No log 1.7451 178 0.7695 0.2974 0.7695 0.8772
No log 1.7647 180 0.5498 0.4455 0.5498 0.7415
No log 1.7843 182 0.4817 0.5358 0.4817 0.6940
No log 1.8039 184 0.4472 0.5495 0.4472 0.6687
No log 1.8235 186 0.4160 0.5757 0.4160 0.6449
No log 1.8431 188 0.4343 0.5719 0.4343 0.6590
No log 1.8627 190 0.4952 0.5948 0.4952 0.7037
No log 1.8824 192 0.5036 0.6279 0.5036 0.7097
No log 1.9020 194 0.5333 0.6159 0.5333 0.7303
No log 1.9216 196 0.6532 0.5700 0.6532 0.8082
No log 1.9412 198 0.7827 0.5312 0.7827 0.8847
No log 1.9608 200 0.7066 0.5591 0.7066 0.8406
No log 1.9804 202 0.5262 0.6230 0.5262 0.7254
No log 2.0 204 0.4205 0.6781 0.4205 0.6485
No log 2.0196 206 0.4190 0.6787 0.4190 0.6473
No log 2.0392 208 0.4111 0.6455 0.4111 0.6412
No log 2.0588 210 0.4902 0.5994 0.4902 0.7002
No log 2.0784 212 0.5573 0.5412 0.5573 0.7465
No log 2.0980 214 0.5682 0.4939 0.5682 0.7538
No log 2.1176 216 0.4962 0.5011 0.4962 0.7044
No log 2.1373 218 0.4108 0.5544 0.4108 0.6409
No log 2.1569 220 0.4203 0.5450 0.4203 0.6483
No log 2.1765 222 0.4189 0.5790 0.4189 0.6472
No log 2.1961 224 0.4279 0.5636 0.4279 0.6542
No log 2.2157 226 0.4954 0.5324 0.4954 0.7038
No log 2.2353 228 0.5269 0.5106 0.5269 0.7259
No log 2.2549 230 0.5232 0.5579 0.5232 0.7233
No log 2.2745 232 0.5288 0.5619 0.5288 0.7272
No log 2.2941 234 0.5143 0.5511 0.5143 0.7172
No log 2.3137 236 0.4644 0.5513 0.4644 0.6815
No log 2.3333 238 0.4138 0.6471 0.4138 0.6432
No log 2.3529 240 0.3884 0.6629 0.3884 0.6232
No log 2.3725 242 0.3927 0.6557 0.3927 0.6267
No log 2.3922 244 0.3944 0.6635 0.3944 0.6280
No log 2.4118 246 0.3940 0.6623 0.3940 0.6277
No log 2.4314 248 0.3832 0.6799 0.3832 0.6190
No log 2.4510 250 0.4090 0.6555 0.4090 0.6395
No log 2.4706 252 0.4466 0.6028 0.4466 0.6683
No log 2.4902 254 0.3929 0.6773 0.3929 0.6268
No log 2.5098 256 0.3974 0.6707 0.3974 0.6304
No log 2.5294 258 0.4757 0.6320 0.4757 0.6897
No log 2.5490 260 0.4944 0.6079 0.4944 0.7032
No log 2.5686 262 0.4317 0.6371 0.4317 0.6570
No log 2.5882 264 0.3744 0.6526 0.3744 0.6119
No log 2.6078 266 0.3651 0.6791 0.3651 0.6043
No log 2.6275 268 0.4125 0.6363 0.4125 0.6423
No log 2.6471 270 0.4578 0.5314 0.4578 0.6766
No log 2.6667 272 0.4430 0.5414 0.4430 0.6656
No log 2.6863 274 0.3968 0.6002 0.3968 0.6299
No log 2.7059 276 0.3961 0.6380 0.3961 0.6293
No log 2.7255 278 0.4105 0.6220 0.4105 0.6407
No log 2.7451 280 0.3911 0.6973 0.3911 0.6254
No log 2.7647 282 0.4421 0.6408 0.4421 0.6649
No log 2.7843 284 0.4349 0.6583 0.4349 0.6595
No log 2.8039 286 0.4395 0.6099 0.4395 0.6629
No log 2.8235 288 0.5627 0.5675 0.5627 0.7501
No log 2.8431 290 0.6906 0.5024 0.6906 0.8310
No log 2.8627 292 0.6768 0.5171 0.6768 0.8227
No log 2.8824 294 0.5395 0.5487 0.5395 0.7345
No log 2.9020 296 0.4011 0.5961 0.4011 0.6333
No log 2.9216 298 0.3767 0.6281 0.3767 0.6138
No log 2.9412 300 0.3731 0.6307 0.3731 0.6108
No log 2.9608 302 0.3733 0.6302 0.3733 0.6110
No log 2.9804 304 0.3923 0.6146 0.3923 0.6263
No log 3.0 306 0.4366 0.5960 0.4366 0.6608
No log 3.0196 308 0.4666 0.5604 0.4666 0.6831
No log 3.0392 310 0.4672 0.5359 0.4672 0.6835
No log 3.0588 312 0.4776 0.5247 0.4776 0.6911
No log 3.0784 314 0.4671 0.5237 0.4671 0.6835
No log 3.0980 316 0.4310 0.5414 0.4310 0.6565
No log 3.1176 318 0.4162 0.6106 0.4162 0.6452
No log 3.1373 320 0.4076 0.6551 0.4076 0.6384
No log 3.1569 322 0.4237 0.6314 0.4237 0.6509
No log 3.1765 324 0.4840 0.5888 0.4840 0.6957
No log 3.1961 326 0.6137 0.4932 0.6137 0.7834
No log 3.2157 328 0.5753 0.4731 0.5753 0.7585
No log 3.2353 330 0.4805 0.6064 0.4805 0.6932
No log 3.2549 332 0.4217 0.6140 0.4217 0.6494
No log 3.2745 334 0.3974 0.6458 0.3974 0.6304
No log 3.2941 336 0.3710 0.6747 0.3710 0.6091
No log 3.3137 338 0.4048 0.6140 0.4048 0.6362
No log 3.3333 340 0.5067 0.5906 0.5067 0.7118
No log 3.3529 342 0.6642 0.5414 0.6642 0.8150
No log 3.3725 344 0.7039 0.5523 0.7039 0.8390
No log 3.3922 346 0.6753 0.5581 0.6753 0.8218
No log 3.4118 348 0.6209 0.5491 0.6209 0.7880
No log 3.4314 350 0.5268 0.5986 0.5268 0.7258
No log 3.4510 352 0.4626 0.6267 0.4626 0.6801
No log 3.4706 354 0.4954 0.6171 0.4954 0.7038
No log 3.4902 356 0.4323 0.6406 0.4323 0.6575
No log 3.5098 358 0.3964 0.6615 0.3964 0.6296
No log 3.5294 360 0.4061 0.6569 0.4061 0.6373
No log 3.5490 362 0.4473 0.6272 0.4473 0.6688
No log 3.5686 364 0.4778 0.6258 0.4778 0.6912
No log 3.5882 366 0.4399 0.6280 0.4399 0.6632
No log 3.6078 368 0.3799 0.6522 0.3799 0.6164
No log 3.6275 370 0.3705 0.6574 0.3705 0.6086
No log 3.6471 372 0.3639 0.6801 0.3639 0.6032
No log 3.6667 374 0.3709 0.6699 0.3709 0.6090
No log 3.6863 376 0.3922 0.6858 0.3922 0.6263
No log 3.7059 378 0.4650 0.6645 0.4650 0.6819
No log 3.7255 380 0.6518 0.5943 0.6518 0.8074
No log 3.7451 382 0.7646 0.5710 0.7646 0.8744
No log 3.7647 384 0.6800 0.5919 0.6800 0.8246
No log 3.7843 386 0.5751 0.6490 0.5751 0.7584
No log 3.8039 388 0.4510 0.6615 0.4510 0.6716
No log 3.8235 390 0.3944 0.6827 0.3944 0.6280
No log 3.8431 392 0.4086 0.7073 0.4086 0.6392
No log 3.8627 394 0.3946 0.6889 0.3946 0.6282
No log 3.8824 396 0.4322 0.6692 0.4322 0.6574
No log 3.9020 398 0.4777 0.6067 0.4777 0.6912
No log 3.9216 400 0.4779 0.5876 0.4779 0.6913
No log 3.9412 402 0.4130 0.6279 0.4130 0.6426
No log 3.9608 404 0.3920 0.6695 0.3920 0.6261
No log 3.9804 406 0.4034 0.6527 0.4034 0.6351
No log 4.0 408 0.4123 0.6516 0.4123 0.6421
No log 4.0196 410 0.4488 0.5796 0.4488 0.6699
No log 4.0392 412 0.5374 0.5288 0.5374 0.7331
No log 4.0588 414 0.5485 0.5675 0.5485 0.7406
No log 4.0784 416 0.5611 0.5623 0.5611 0.7490
No log 4.0980 418 0.5286 0.5926 0.5286 0.7271
No log 4.1176 420 0.4814 0.6253 0.4814 0.6938
No log 4.1373 422 0.4475 0.6719 0.4475 0.6690
No log 4.1569 424 0.4315 0.6881 0.4315 0.6569
No log 4.1765 426 0.4047 0.6739 0.4047 0.6361
No log 4.1961 428 0.4295 0.6430 0.4295 0.6554
No log 4.2157 430 0.4651 0.6224 0.4651 0.6820
No log 4.2353 432 0.4235 0.6340 0.4235 0.6508
No log 4.2549 434 0.3942 0.6447 0.3942 0.6278
No log 4.2745 436 0.3982 0.6489 0.3982 0.6310
No log 4.2941 438 0.4764 0.5611 0.4764 0.6903
No log 4.3137 440 0.5092 0.5448 0.5092 0.7136
No log 4.3333 442 0.4436 0.5924 0.4436 0.6660
No log 4.3529 444 0.3878 0.6386 0.3878 0.6227
No log 4.3725 446 0.3776 0.6462 0.3776 0.6145
No log 4.3922 448 0.4046 0.6441 0.4046 0.6361
No log 4.4118 450 0.4609 0.6176 0.4609 0.6789
No log 4.4314 452 0.5155 0.6013 0.5155 0.7180
No log 4.4510 454 0.4932 0.6299 0.4932 0.7023
No log 4.4706 456 0.4804 0.6326 0.4804 0.6931
No log 4.4902 458 0.4366 0.6672 0.4366 0.6607
No log 4.5098 460 0.4140 0.6884 0.4140 0.6434
No log 4.5294 462 0.4564 0.6646 0.4564 0.6756
No log 4.5490 464 0.4848 0.6457 0.4848 0.6963
No log 4.5686 466 0.4989 0.6308 0.4989 0.7063
No log 4.5882 468 0.4408 0.6615 0.4408 0.6640
No log 4.6078 470 0.4229 0.6560 0.4229 0.6503
No log 4.6275 472 0.3739 0.6844 0.3739 0.6115
No log 4.6471 474 0.3633 0.6968 0.3633 0.6028
No log 4.6667 476 0.3928 0.6647 0.3928 0.6267
No log 4.6863 478 0.4538 0.6462 0.4538 0.6737
No log 4.7059 480 0.5378 0.5818 0.5378 0.7334
No log 4.7255 482 0.5477 0.5750 0.5477 0.7401
No log 4.7451 484 0.5435 0.5955 0.5435 0.7372
No log 4.7647 486 0.4984 0.6274 0.4984 0.7059
No log 4.7843 488 0.4782 0.6469 0.4782 0.6915
No log 4.8039 490 0.4183 0.6771 0.4183 0.6467
No log 4.8235 492 0.4049 0.7001 0.4049 0.6363
No log 4.8431 494 0.4742 0.6781 0.4742 0.6886
No log 4.8627 496 0.5400 0.6232 0.5400 0.7348
No log 4.8824 498 0.5069 0.6391 0.5069 0.7120
0.4786 4.9020 500 0.4636 0.6511 0.4636 0.6809
0.4786 4.9216 502 0.4543 0.6447 0.4543 0.6740
0.4786 4.9412 504 0.4543 0.6353 0.4543 0.6740
0.4786 4.9608 506 0.4914 0.5705 0.4914 0.7010
0.4786 4.9804 508 0.4954 0.5341 0.4954 0.7039
0.4786 5.0 510 0.4744 0.5496 0.4744 0.6887

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask8_development

Finetuned
(4019)
this model