Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask4_style

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4517
  • Qwk: 0.6452
  • Mse: 0.4517
  • Rmse: 0.6721

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0187 2 4.5724 0.0052 4.5724 2.1383
No log 0.0374 4 3.2799 0.0572 3.2799 1.8111
No log 0.0561 6 1.7337 0.1198 1.7337 1.3167
No log 0.0748 8 0.9329 0.1454 0.9329 0.9659
No log 0.0935 10 0.8469 0.1777 0.8469 0.9203
No log 0.1121 12 0.8422 0.1185 0.8422 0.9177
No log 0.1308 14 0.8892 0.1323 0.8892 0.9430
No log 0.1495 16 0.8657 0.1463 0.8657 0.9304
No log 0.1682 18 0.8127 0.2115 0.8127 0.9015
No log 0.1869 20 0.7497 0.2037 0.7497 0.8659
No log 0.2056 22 0.6409 0.3286 0.6409 0.8006
No log 0.2243 24 0.6263 0.3744 0.6263 0.7914
No log 0.2430 26 0.6386 0.4394 0.6386 0.7991
No log 0.2617 28 0.6057 0.4083 0.6057 0.7783
No log 0.2804 30 0.6045 0.3670 0.6045 0.7775
No log 0.2991 32 0.6473 0.3126 0.6473 0.8045
No log 0.3178 34 0.6536 0.3038 0.6536 0.8084
No log 0.3364 36 0.6845 0.2931 0.6845 0.8273
No log 0.3551 38 0.6419 0.2884 0.6419 0.8012
No log 0.3738 40 0.6578 0.2791 0.6578 0.8110
No log 0.3925 42 0.6052 0.3100 0.6052 0.7779
No log 0.4112 44 0.5540 0.3578 0.5540 0.7443
No log 0.4299 46 0.5837 0.3666 0.5837 0.7640
No log 0.4486 48 0.6072 0.3650 0.6072 0.7793
No log 0.4673 50 0.5897 0.3845 0.5897 0.7679
No log 0.4860 52 0.5447 0.4350 0.5447 0.7380
No log 0.5047 54 0.4780 0.5102 0.4780 0.6914
No log 0.5234 56 0.4895 0.5333 0.4895 0.6997
No log 0.5421 58 0.5694 0.4410 0.5694 0.7546
No log 0.5607 60 0.7398 0.2825 0.7398 0.8601
No log 0.5794 62 0.7281 0.3123 0.7281 0.8533
No log 0.5981 64 0.6619 0.4145 0.6619 0.8136
No log 0.6168 66 0.5573 0.5355 0.5573 0.7465
No log 0.6355 68 0.4412 0.6559 0.4412 0.6642
No log 0.6542 70 0.4761 0.6260 0.4761 0.6900
No log 0.6729 72 0.4425 0.6264 0.4425 0.6652
No log 0.6916 74 0.5704 0.5275 0.5704 0.7552
No log 0.7103 76 0.5444 0.5379 0.5444 0.7378
No log 0.7290 78 0.4688 0.5552 0.4688 0.6847
No log 0.7477 80 0.4341 0.5795 0.4341 0.6589
No log 0.7664 82 0.4273 0.6428 0.4273 0.6537
No log 0.7850 84 0.4267 0.6568 0.4267 0.6532
No log 0.8037 86 0.4348 0.6440 0.4348 0.6594
No log 0.8224 88 0.4838 0.6055 0.4838 0.6956
No log 0.8411 90 0.5436 0.5642 0.5436 0.7373
No log 0.8598 92 0.4848 0.5135 0.4848 0.6963
No log 0.8785 94 0.4961 0.4662 0.4961 0.7044
No log 0.8972 96 0.4977 0.4851 0.4977 0.7055
No log 0.9159 98 0.5120 0.5425 0.5120 0.7156
No log 0.9346 100 0.5100 0.5823 0.5100 0.7141
No log 0.9533 102 0.5730 0.5320 0.5730 0.7570
No log 0.9720 104 0.5533 0.5578 0.5533 0.7438
No log 0.9907 106 0.4788 0.6297 0.4788 0.6920
No log 1.0093 108 0.4702 0.6478 0.4702 0.6857
No log 1.0280 110 0.4772 0.6494 0.4772 0.6908
No log 1.0467 112 0.5219 0.5850 0.5219 0.7224
No log 1.0654 114 0.5688 0.5613 0.5688 0.7542
No log 1.0841 116 0.5747 0.5700 0.5747 0.7581
No log 1.1028 118 0.5495 0.5955 0.5495 0.7413
No log 1.1215 120 0.4697 0.5977 0.4697 0.6854
No log 1.1402 122 0.5303 0.5984 0.5303 0.7282
No log 1.1589 124 0.4965 0.5778 0.4965 0.7046
No log 1.1776 126 0.4938 0.5765 0.4938 0.7027
No log 1.1963 128 0.4785 0.5771 0.4785 0.6918
No log 1.2150 130 0.4733 0.5578 0.4733 0.6880
No log 1.2336 132 0.4528 0.6133 0.4528 0.6729
No log 1.2523 134 0.4471 0.6754 0.4471 0.6686
No log 1.2710 136 0.5303 0.6399 0.5303 0.7282
No log 1.2897 138 0.6858 0.5521 0.6858 0.8281
No log 1.3084 140 0.5764 0.5937 0.5764 0.7592
No log 1.3271 142 0.4902 0.6664 0.4902 0.7001
No log 1.3458 144 0.4242 0.6901 0.4242 0.6513
No log 1.3645 146 0.4500 0.6575 0.4500 0.6708
No log 1.3832 148 0.4653 0.6223 0.4653 0.6821
No log 1.4019 150 0.4806 0.5852 0.4806 0.6933
No log 1.4206 152 0.4573 0.5749 0.4573 0.6762
No log 1.4393 154 0.5137 0.3729 0.5137 0.7167
No log 1.4579 156 0.5948 0.3607 0.5948 0.7712
No log 1.4766 158 0.6425 0.4892 0.6425 0.8016
No log 1.4953 160 0.5874 0.5019 0.5874 0.7664
No log 1.5140 162 0.4400 0.5626 0.4400 0.6633
No log 1.5327 164 0.3889 0.6453 0.3889 0.6236
No log 1.5514 166 0.4045 0.6893 0.4045 0.6360
No log 1.5701 168 0.4126 0.7298 0.4126 0.6424
No log 1.5888 170 0.5553 0.6554 0.5553 0.7452
No log 1.6075 172 0.7063 0.5398 0.7063 0.8404
No log 1.6262 174 0.7279 0.3983 0.7279 0.8531
No log 1.6449 176 0.6280 0.4095 0.6280 0.7925
No log 1.6636 178 0.4817 0.5323 0.4817 0.6941
No log 1.6822 180 0.4209 0.6529 0.4209 0.6488
No log 1.7009 182 0.4305 0.6549 0.4305 0.6561
No log 1.7196 184 0.5500 0.5554 0.5500 0.7416
No log 1.7383 186 0.5718 0.5512 0.5718 0.7562
No log 1.7570 188 0.4614 0.6040 0.4614 0.6793
No log 1.7757 190 0.4231 0.6312 0.4231 0.6505
No log 1.7944 192 0.4226 0.6250 0.4226 0.6500
No log 1.8131 194 0.4085 0.6545 0.4085 0.6391
No log 1.8318 196 0.4014 0.6730 0.4014 0.6336
No log 1.8505 198 0.4605 0.6447 0.4605 0.6786
No log 1.8692 200 0.5895 0.5825 0.5895 0.7678
No log 1.8879 202 0.5487 0.6286 0.5487 0.7407
No log 1.9065 204 0.4204 0.6867 0.4204 0.6484
No log 1.9252 206 0.4056 0.6809 0.4056 0.6369
No log 1.9439 208 0.4251 0.6805 0.4251 0.6520
No log 1.9626 210 0.5432 0.6058 0.5432 0.7370
No log 1.9813 212 0.6024 0.5872 0.6024 0.7761
No log 2.0 214 0.5820 0.6060 0.5820 0.7629
No log 2.0187 216 0.5252 0.6505 0.5252 0.7247
No log 2.0374 218 0.5200 0.6708 0.5200 0.7211
No log 2.0561 220 0.4722 0.6990 0.4722 0.6872
No log 2.0748 222 0.4148 0.6760 0.4148 0.6441
No log 2.0935 224 0.4234 0.6883 0.4234 0.6507
No log 2.1121 226 0.4236 0.7008 0.4236 0.6508
No log 2.1308 228 0.4907 0.6798 0.4907 0.7005
No log 2.1495 230 0.4485 0.6981 0.4485 0.6697
No log 2.1682 232 0.3970 0.6846 0.3970 0.6301
No log 2.1869 234 0.3895 0.6807 0.3895 0.6241
No log 2.2056 236 0.3913 0.6725 0.3913 0.6256
No log 2.2243 238 0.3938 0.6550 0.3938 0.6275
No log 2.2430 240 0.4173 0.6408 0.4173 0.6460
No log 2.2617 242 0.4816 0.6096 0.4816 0.6940
No log 2.2804 244 0.4640 0.5896 0.4640 0.6811
No log 2.2991 246 0.4401 0.6321 0.4401 0.6634
No log 2.3178 248 0.4277 0.6732 0.4277 0.6540
No log 2.3364 250 0.4778 0.6806 0.4778 0.6913
No log 2.3551 252 0.6617 0.5923 0.6617 0.8135
No log 2.3738 254 0.7443 0.5645 0.7443 0.8627
No log 2.3925 256 0.5647 0.6490 0.5647 0.7515
No log 2.4112 258 0.4355 0.7124 0.4355 0.6599
No log 2.4299 260 0.4403 0.7136 0.4403 0.6635
No log 2.4486 262 0.4550 0.7243 0.4550 0.6745
No log 2.4673 264 0.4518 0.7149 0.4518 0.6721
No log 2.4860 266 0.4470 0.6996 0.4470 0.6686
No log 2.5047 268 0.5201 0.6754 0.5201 0.7212
No log 2.5234 270 0.5257 0.6603 0.5257 0.7250
No log 2.5421 272 0.4159 0.6433 0.4159 0.6449
No log 2.5607 274 0.4120 0.5960 0.4120 0.6419
No log 2.5794 276 0.4890 0.6095 0.4890 0.6993
No log 2.5981 278 0.5225 0.5924 0.5225 0.7229
No log 2.6168 280 0.4565 0.5826 0.4565 0.6757
No log 2.6355 282 0.4149 0.6198 0.4149 0.6441
No log 2.6542 284 0.4185 0.6104 0.4185 0.6469
No log 2.6729 286 0.4826 0.5636 0.4826 0.6947
No log 2.6916 288 0.6556 0.5229 0.6556 0.8097
No log 2.7103 290 0.7133 0.5226 0.7133 0.8446
No log 2.7290 292 0.6149 0.5340 0.6149 0.7841
No log 2.7477 294 0.5276 0.6002 0.5276 0.7263
No log 2.7664 296 0.4741 0.6178 0.4741 0.6886
No log 2.7850 298 0.4516 0.6569 0.4516 0.6720
No log 2.8037 300 0.4581 0.6458 0.4581 0.6769
No log 2.8224 302 0.5012 0.6409 0.5012 0.7079
No log 2.8411 304 0.6311 0.6055 0.6311 0.7944
No log 2.8598 306 0.7749 0.5577 0.7749 0.8803
No log 2.8785 308 0.6747 0.5740 0.6747 0.8214
No log 2.8972 310 0.4672 0.6208 0.4672 0.6835
No log 2.9159 312 0.4050 0.6802 0.4050 0.6364
No log 2.9346 314 0.4126 0.6934 0.4126 0.6423
No log 2.9533 316 0.4261 0.6845 0.4261 0.6528
No log 2.9720 318 0.4154 0.6800 0.4154 0.6445
No log 2.9907 320 0.5348 0.6124 0.5348 0.7313
No log 3.0093 322 0.7078 0.5423 0.7078 0.8413
No log 3.0280 324 0.6585 0.5662 0.6585 0.8115
No log 3.0467 326 0.5156 0.5300 0.5156 0.7180
No log 3.0654 328 0.4817 0.5709 0.4817 0.6940
No log 3.0841 330 0.4906 0.5642 0.4906 0.7004
No log 3.1028 332 0.4801 0.5745 0.4801 0.6929
No log 3.1215 334 0.4860 0.5583 0.4860 0.6971
No log 3.1402 336 0.5077 0.5794 0.5077 0.7125
No log 3.1589 338 0.4782 0.6269 0.4782 0.6915
No log 3.1776 340 0.4226 0.6481 0.4226 0.6501
No log 3.1963 342 0.4271 0.6679 0.4271 0.6535
No log 3.2150 344 0.4757 0.6355 0.4757 0.6897
No log 3.2336 346 0.4988 0.6312 0.4988 0.7063
No log 3.2523 348 0.5596 0.6470 0.5596 0.7481
No log 3.2710 350 0.6296 0.6296 0.6296 0.7935
No log 3.2897 352 0.5166 0.6847 0.5166 0.7187
No log 3.3084 354 0.4798 0.6737 0.4798 0.6927
No log 3.3271 356 0.4650 0.6711 0.4650 0.6819
No log 3.3458 358 0.5791 0.6481 0.5791 0.7610
No log 3.3645 360 0.8428 0.4936 0.8428 0.9180
No log 3.3832 362 0.8395 0.4929 0.8395 0.9163
No log 3.4019 364 0.6271 0.5887 0.6271 0.7919
No log 3.4206 366 0.4691 0.6435 0.4691 0.6849
No log 3.4393 368 0.4627 0.6349 0.4627 0.6803
No log 3.4579 370 0.4620 0.6354 0.4620 0.6797
No log 3.4766 372 0.5638 0.6098 0.5638 0.7509
No log 3.4953 374 0.6931 0.5360 0.6931 0.8325
No log 3.5140 376 0.6538 0.5698 0.6538 0.8086
No log 3.5327 378 0.4998 0.6713 0.4998 0.7070
No log 3.5514 380 0.4314 0.7049 0.4314 0.6568
No log 3.5701 382 0.4330 0.7199 0.4330 0.6581
No log 3.5888 384 0.4393 0.7212 0.4393 0.6628
No log 3.6075 386 0.4841 0.7141 0.4841 0.6957
No log 3.6262 388 0.5707 0.6736 0.5707 0.7554
No log 3.6449 390 0.6862 0.6074 0.6862 0.8284
No log 3.6636 392 0.6814 0.5984 0.6814 0.8255
No log 3.6822 394 0.5064 0.6594 0.5064 0.7116
No log 3.7009 396 0.4073 0.6720 0.4073 0.6382
No log 3.7196 398 0.4399 0.6440 0.4399 0.6633
No log 3.7383 400 0.4312 0.6786 0.4312 0.6566
No log 3.7570 402 0.4201 0.6960 0.4201 0.6481
No log 3.7757 404 0.4755 0.6421 0.4755 0.6896
No log 3.7944 406 0.6258 0.5975 0.6258 0.7911
No log 3.8131 408 0.7476 0.5548 0.7476 0.8647
No log 3.8318 410 0.6933 0.5723 0.6933 0.8326
No log 3.8505 412 0.5060 0.5698 0.5060 0.7113
No log 3.8692 414 0.4109 0.6062 0.4109 0.6411
No log 3.8879 416 0.4067 0.6373 0.4067 0.6377
No log 3.9065 418 0.3965 0.6396 0.3965 0.6297
No log 3.9252 420 0.4304 0.6655 0.4304 0.6560
No log 3.9439 422 0.5460 0.6062 0.5460 0.7389
No log 3.9626 424 0.6663 0.5852 0.6663 0.8163
No log 3.9813 426 0.5943 0.6208 0.5943 0.7709
No log 4.0 428 0.4460 0.6838 0.4460 0.6678
No log 4.0187 430 0.4194 0.7057 0.4194 0.6476
No log 4.0374 432 0.3993 0.7161 0.3993 0.6319
No log 4.0561 434 0.4055 0.6440 0.4055 0.6368
No log 4.0748 436 0.5448 0.5762 0.5448 0.7381
No log 4.0935 438 0.6455 0.5322 0.6455 0.8034
No log 4.1121 440 0.6322 0.5526 0.6322 0.7951
No log 4.1308 442 0.5559 0.5840 0.5559 0.7456
No log 4.1495 444 0.4345 0.6179 0.4345 0.6592
No log 4.1682 446 0.4190 0.6585 0.4190 0.6473
No log 4.1869 448 0.4135 0.6866 0.4135 0.6430
No log 4.2056 450 0.4520 0.6461 0.4520 0.6723
No log 4.2243 452 0.4902 0.6383 0.4902 0.7002
No log 4.2430 454 0.4884 0.6690 0.4884 0.6988
No log 4.2617 456 0.5232 0.6464 0.5232 0.7233
No log 4.2804 458 0.5055 0.6500 0.5055 0.7110
No log 4.2991 460 0.5262 0.6610 0.5262 0.7254
No log 4.3178 462 0.4769 0.6632 0.4769 0.6906
No log 4.3364 464 0.4352 0.6530 0.4352 0.6597
No log 4.3551 466 0.4351 0.6461 0.4351 0.6596
No log 4.3738 468 0.5413 0.6775 0.5413 0.7357
No log 4.3925 470 0.7915 0.5601 0.7915 0.8897
No log 4.4112 472 0.8316 0.5368 0.8316 0.9119
No log 4.4299 474 0.6450 0.5955 0.6450 0.8031
No log 4.4486 476 0.5217 0.6697 0.5217 0.7223
No log 4.4673 478 0.4405 0.6664 0.4405 0.6637
No log 4.4860 480 0.4380 0.6594 0.4380 0.6618
No log 4.5047 482 0.4122 0.6525 0.4122 0.6420
No log 4.5234 484 0.4002 0.6340 0.4002 0.6326
No log 4.5421 486 0.4024 0.6290 0.4024 0.6344
No log 4.5607 488 0.4541 0.6410 0.4541 0.6738
No log 4.5794 490 0.5284 0.5939 0.5284 0.7269
No log 4.5981 492 0.5348 0.6014 0.5348 0.7313
No log 4.6168 494 0.6401 0.5780 0.6401 0.8001
No log 4.6355 496 0.5358 0.6459 0.5358 0.7320
No log 4.6542 498 0.4410 0.6360 0.4410 0.6641
0.4693 4.6729 500 0.4715 0.6597 0.4715 0.6866
0.4693 4.6916 502 0.6058 0.6460 0.6058 0.7783
0.4693 4.7103 504 0.5520 0.6582 0.5520 0.7429
0.4693 4.7290 506 0.4763 0.6889 0.4763 0.6902
0.4693 4.7477 508 0.4163 0.7048 0.4163 0.6452
0.4693 4.7664 510 0.4130 0.6856 0.4130 0.6427
0.4693 4.7850 512 0.4526 0.6749 0.4526 0.6728
0.4693 4.8037 514 0.5144 0.6490 0.5144 0.7172
0.4693 4.8224 516 0.5604 0.6067 0.5604 0.7486
0.4693 4.8411 518 0.5533 0.6267 0.5533 0.7439
0.4693 4.8598 520 0.4746 0.6688 0.4746 0.6889
0.4693 4.8785 522 0.4421 0.6570 0.4421 0.6649
0.4693 4.8972 524 0.4153 0.6330 0.4153 0.6444
0.4693 4.9159 526 0.4519 0.6482 0.4519 0.6722
0.4693 4.9346 528 0.4574 0.6552 0.4574 0.6763
0.4693 4.9533 530 0.4612 0.6639 0.4612 0.6791
0.4693 4.9720 532 0.4031 0.6971 0.4031 0.6349
0.4693 4.9907 534 0.4050 0.7243 0.4050 0.6364
0.4693 5.0093 536 0.4836 0.6867 0.4836 0.6954
0.4693 5.0280 538 0.5560 0.6670 0.5560 0.7456
0.4693 5.0467 540 0.4559 0.6934 0.4559 0.6752
0.4693 5.0654 542 0.4226 0.6868 0.4226 0.6501
0.4693 5.0841 544 0.4591 0.6467 0.4591 0.6776
0.4693 5.1028 546 0.4498 0.6382 0.4498 0.6707
0.4693 5.1215 548 0.4588 0.6260 0.4588 0.6773
0.4693 5.1402 550 0.4780 0.6404 0.4780 0.6914
0.4693 5.1589 552 0.4517 0.6452 0.4517 0.6721

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/Arabic_CrossPrompt_FineTuningAraBERT_noAug_TestTask4_style

Finetuned
(4019)
this model