ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k14_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4929
  • Qwk: 0.4847
  • Mse: 0.4929
  • Rmse: 0.7021

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 2.5519 -0.1213 2.5519 1.5975
No log 0.1111 4 1.3000 0.0412 1.3000 1.1402
No log 0.1667 6 1.1046 0.0476 1.1046 1.0510
No log 0.2222 8 0.8966 0.0408 0.8966 0.9469
No log 0.2778 10 0.8234 0.1093 0.8234 0.9074
No log 0.3333 12 0.9243 0.0058 0.9243 0.9614
No log 0.3889 14 0.9262 0.0478 0.9262 0.9624
No log 0.4444 16 0.8842 0.0717 0.8842 0.9403
No log 0.5 18 0.8177 0.0643 0.8177 0.9043
No log 0.5556 20 0.7688 0.1187 0.7688 0.8768
No log 0.6111 22 0.7551 0.0428 0.7551 0.8690
No log 0.6667 24 0.7611 0.0428 0.7611 0.8724
No log 0.7222 26 0.7364 0.0840 0.7364 0.8581
No log 0.7778 28 0.7772 0.2883 0.7772 0.8816
No log 0.8333 30 0.8088 0.3105 0.8088 0.8993
No log 0.8889 32 0.7655 0.2692 0.7655 0.8749
No log 0.9444 34 0.7013 0.3575 0.7013 0.8374
No log 1.0 36 0.7439 0.2317 0.7439 0.8625
No log 1.0556 38 0.7078 0.2783 0.7078 0.8413
No log 1.1111 40 0.7391 0.3637 0.7391 0.8597
No log 1.1667 42 0.8616 0.2843 0.8616 0.9282
No log 1.2222 44 0.8755 0.1962 0.8755 0.9357
No log 1.2778 46 0.8959 0.2923 0.8959 0.9465
No log 1.3333 48 0.8697 0.3425 0.8697 0.9326
No log 1.3889 50 0.7804 0.3238 0.7804 0.8834
No log 1.4444 52 0.7288 0.2285 0.7288 0.8537
No log 1.5 54 0.6840 0.2718 0.6840 0.8270
No log 1.5556 56 0.6796 0.2652 0.6796 0.8244
No log 1.6111 58 0.6982 0.2227 0.6982 0.8356
No log 1.6667 60 0.7411 0.2527 0.7411 0.8608
No log 1.7222 62 0.7601 0.2817 0.7601 0.8718
No log 1.7778 64 0.7568 0.2817 0.7568 0.8699
No log 1.8333 66 0.8656 0.2615 0.8656 0.9304
No log 1.8889 68 0.8814 0.2615 0.8814 0.9388
No log 1.9444 70 0.8785 0.2995 0.8785 0.9373
No log 2.0 72 0.7780 0.1866 0.7780 0.8820
No log 2.0556 74 0.7609 0.2171 0.7609 0.8723
No log 2.1111 76 0.7895 0.2518 0.7895 0.8886
No log 2.1667 78 0.7907 0.3231 0.7907 0.8892
No log 2.2222 80 0.7271 0.1962 0.7271 0.8527
No log 2.2778 82 0.7466 0.3069 0.7466 0.8641
No log 2.3333 84 0.6717 0.3399 0.6717 0.8196
No log 2.3889 86 0.6348 0.3426 0.6348 0.7967
No log 2.4444 88 0.6295 0.4068 0.6295 0.7934
No log 2.5 90 0.6110 0.4068 0.6110 0.7817
No log 2.5556 92 0.6533 0.3919 0.6533 0.8083
No log 2.6111 94 0.6640 0.4437 0.6640 0.8149
No log 2.6667 96 0.6156 0.3523 0.6156 0.7846
No log 2.7222 98 0.6404 0.3728 0.6404 0.8003
No log 2.7778 100 0.8311 0.2467 0.8311 0.9116
No log 2.8333 102 0.7053 0.3224 0.7053 0.8398
No log 2.8889 104 0.6530 0.2218 0.6530 0.8081
No log 2.9444 106 0.6970 0.2193 0.6970 0.8348
No log 3.0 108 0.9156 0.1672 0.9156 0.9569
No log 3.0556 110 0.8569 0.2443 0.8569 0.9257
No log 3.1111 112 0.5991 0.4618 0.5991 0.7740
No log 3.1667 114 0.5557 0.5266 0.5557 0.7454
No log 3.2222 116 0.5431 0.5248 0.5431 0.7370
No log 3.2778 118 0.5384 0.4829 0.5384 0.7338
No log 3.3333 120 0.5472 0.4595 0.5472 0.7397
No log 3.3889 122 0.5403 0.4555 0.5403 0.7350
No log 3.4444 124 0.5466 0.5430 0.5466 0.7393
No log 3.5 126 0.6177 0.4704 0.6177 0.7859
No log 3.5556 128 0.8762 0.3044 0.8762 0.9361
No log 3.6111 130 1.1158 0.1967 1.1158 1.0563
No log 3.6667 132 0.9341 0.2988 0.9341 0.9665
No log 3.7222 134 0.5717 0.5022 0.5717 0.7561
No log 3.7778 136 0.5997 0.4171 0.5997 0.7744
No log 3.8333 138 0.5736 0.4990 0.5736 0.7574
No log 3.8889 140 0.6469 0.4100 0.6469 0.8043
No log 3.9444 142 0.7402 0.3653 0.7402 0.8604
No log 4.0 144 0.6150 0.3894 0.6150 0.7842
No log 4.0556 146 0.5404 0.5521 0.5404 0.7351
No log 4.1111 148 0.5489 0.5404 0.5489 0.7409
No log 4.1667 150 0.5399 0.5326 0.5399 0.7347
No log 4.2222 152 0.5423 0.5095 0.5423 0.7364
No log 4.2778 154 0.5221 0.4788 0.5221 0.7226
No log 4.3333 156 0.5160 0.5714 0.5160 0.7184
No log 4.3889 158 0.5097 0.5798 0.5097 0.7140
No log 4.4444 160 0.4981 0.6154 0.4981 0.7057
No log 4.5 162 0.5262 0.4858 0.5262 0.7254
No log 4.5556 164 0.5771 0.4971 0.5771 0.7597
No log 4.6111 166 0.5111 0.5267 0.5111 0.7149
No log 4.6667 168 0.4911 0.5812 0.4911 0.7008
No log 4.7222 170 0.5096 0.6634 0.5096 0.7138
No log 4.7778 172 0.5105 0.5782 0.5105 0.7145
No log 4.8333 174 0.5258 0.5286 0.5258 0.7251
No log 4.8889 176 0.5373 0.5352 0.5373 0.7330
No log 4.9444 178 0.5428 0.6009 0.5428 0.7367
No log 5.0 180 0.6498 0.6016 0.6498 0.8061
No log 5.0556 182 0.5940 0.5668 0.5940 0.7707
No log 5.1111 184 0.5362 0.5860 0.5362 0.7322
No log 5.1667 186 0.5520 0.4942 0.5520 0.7429
No log 5.2222 188 0.5340 0.5860 0.5340 0.7307
No log 5.2778 190 0.6863 0.4072 0.6863 0.8285
No log 5.3333 192 0.7453 0.3699 0.7453 0.8633
No log 5.3889 194 0.6728 0.3590 0.6728 0.8202
No log 5.4444 196 0.6384 0.3996 0.6384 0.7990
No log 5.5 198 0.5975 0.4270 0.5975 0.7730
No log 5.5556 200 0.5908 0.4835 0.5908 0.7687
No log 5.6111 202 0.7368 0.4251 0.7368 0.8583
No log 5.6667 204 0.7528 0.4180 0.7528 0.8677
No log 5.7222 206 0.7578 0.4114 0.7578 0.8705
No log 5.7778 208 0.5757 0.4911 0.5757 0.7588
No log 5.8333 210 0.4877 0.5533 0.4877 0.6983
No log 5.8889 212 0.4803 0.5941 0.4803 0.6931
No log 5.9444 214 0.4784 0.5609 0.4784 0.6917
No log 6.0 216 0.5447 0.4576 0.5447 0.7380
No log 6.0556 218 0.5705 0.4576 0.5705 0.7553
No log 6.1111 220 0.4956 0.5736 0.4956 0.7040
No log 6.1667 222 0.4721 0.6129 0.4721 0.6871
No log 6.2222 224 0.4621 0.6254 0.4621 0.6798
No log 6.2778 226 0.4679 0.5980 0.4679 0.6840
No log 6.3333 228 0.4502 0.6254 0.4502 0.6710
No log 6.3889 230 0.4482 0.6254 0.4482 0.6695
No log 6.4444 232 0.4658 0.6346 0.4658 0.6825
No log 6.5 234 0.4665 0.6443 0.4665 0.6830
No log 6.5556 236 0.5344 0.5787 0.5344 0.7311
No log 6.6111 238 0.6781 0.4805 0.6781 0.8235
No log 6.6667 240 0.6973 0.4801 0.6973 0.8350
No log 6.7222 242 0.5906 0.4997 0.5906 0.7685
No log 6.7778 244 0.6019 0.4905 0.6019 0.7758
No log 6.8333 246 0.6567 0.4491 0.6567 0.8104
No log 6.8889 248 0.5952 0.4997 0.5952 0.7715
No log 6.9444 250 0.5464 0.4229 0.5464 0.7392
No log 7.0 252 0.5333 0.4217 0.5333 0.7303
No log 7.0556 254 0.5343 0.4338 0.5343 0.7310
No log 7.1111 256 0.5963 0.5017 0.5963 0.7722
No log 7.1667 258 0.6373 0.4330 0.6373 0.7983
No log 7.2222 260 0.5698 0.4614 0.5698 0.7549
No log 7.2778 262 0.5390 0.4378 0.5390 0.7341
No log 7.3333 264 0.5275 0.4752 0.5275 0.7263
No log 7.3889 266 0.5702 0.5639 0.5702 0.7551
No log 7.4444 268 0.6010 0.5544 0.6010 0.7753
No log 7.5 270 0.5426 0.5597 0.5426 0.7366
No log 7.5556 272 0.5020 0.5584 0.5020 0.7085
No log 7.6111 274 0.5021 0.5305 0.5021 0.7086
No log 7.6667 276 0.5077 0.4402 0.5077 0.7125
No log 7.7222 278 0.5115 0.5151 0.5115 0.7152
No log 7.7778 280 0.5069 0.4898 0.5069 0.7120
No log 7.8333 282 0.5029 0.5339 0.5029 0.7091
No log 7.8889 284 0.4869 0.6317 0.4869 0.6977
No log 7.9444 286 0.6217 0.5026 0.6217 0.7885
No log 8.0 288 0.7558 0.3963 0.7558 0.8694
No log 8.0556 290 0.6495 0.4349 0.6495 0.8059
No log 8.1111 292 0.4947 0.5715 0.4947 0.7034
No log 8.1667 294 0.4724 0.5714 0.4724 0.6873
No log 8.2222 296 0.4683 0.5960 0.4683 0.6843
No log 8.2778 298 0.5004 0.5309 0.5004 0.7074
No log 8.3333 300 0.4768 0.6037 0.4768 0.6905
No log 8.3889 302 0.4831 0.5485 0.4831 0.6951
No log 8.4444 304 0.5606 0.5252 0.5606 0.7487
No log 8.5 306 0.6434 0.4728 0.6434 0.8021
No log 8.5556 308 0.7144 0.4648 0.7144 0.8452
No log 8.6111 310 0.7184 0.4721 0.7184 0.8476
No log 8.6667 312 0.5966 0.4916 0.5966 0.7724
No log 8.7222 314 0.5623 0.5723 0.5623 0.7499
No log 8.7778 316 0.6113 0.5140 0.6113 0.7818
No log 8.8333 318 0.7170 0.5665 0.7170 0.8467
No log 8.8889 320 0.6600 0.5594 0.6600 0.8124
No log 8.9444 322 0.5506 0.5692 0.5506 0.7420
No log 9.0 324 0.4969 0.5095 0.4969 0.7049
No log 9.0556 326 0.4816 0.5672 0.4816 0.6939
No log 9.1111 328 0.4765 0.5114 0.4765 0.6903
No log 9.1667 330 0.5160 0.4795 0.5160 0.7183
No log 9.2222 332 0.5463 0.5367 0.5463 0.7391
No log 9.2778 334 0.5030 0.4206 0.5030 0.7092
No log 9.3333 336 0.4945 0.5304 0.4945 0.7032
No log 9.3889 338 0.5324 0.5603 0.5324 0.7296
No log 9.4444 340 0.5207 0.5304 0.5207 0.7216
No log 9.5 342 0.5168 0.5095 0.5168 0.7189
No log 9.5556 344 0.5383 0.3919 0.5383 0.7337
No log 9.6111 346 0.5499 0.3919 0.5499 0.7415
No log 9.6667 348 0.5320 0.4001 0.5320 0.7294
No log 9.7222 350 0.5213 0.4378 0.5213 0.7220
No log 9.7778 352 0.5060 0.4929 0.5060 0.7113
No log 9.8333 354 0.4964 0.5114 0.4964 0.7045
No log 9.8889 356 0.5246 0.4397 0.5246 0.7243
No log 9.9444 358 0.5884 0.5622 0.5884 0.7671
No log 10.0 360 0.5373 0.5404 0.5373 0.7330
No log 10.0556 362 0.5039 0.5557 0.5039 0.7099
No log 10.1111 364 0.5041 0.5353 0.5041 0.7100
No log 10.1667 366 0.4941 0.4919 0.4941 0.7030
No log 10.2222 368 0.5046 0.4808 0.5046 0.7103
No log 10.2778 370 0.5106 0.3919 0.5106 0.7146
No log 10.3333 372 0.4998 0.5101 0.4998 0.7070
No log 10.3889 374 0.4904 0.5286 0.4904 0.7003
No log 10.4444 376 0.4835 0.4970 0.4835 0.6954
No log 10.5 378 0.4807 0.5731 0.4807 0.6933
No log 10.5556 380 0.4682 0.5044 0.4682 0.6842
No log 10.6111 382 0.4910 0.4482 0.4910 0.7007
No log 10.6667 384 0.5073 0.3919 0.5073 0.7123
No log 10.7222 386 0.4763 0.5101 0.4763 0.6901
No log 10.7778 388 0.4824 0.5305 0.4824 0.6945
No log 10.8333 390 0.5022 0.5162 0.5022 0.7087
No log 10.8889 392 0.5045 0.4774 0.5045 0.7103
No log 10.9444 394 0.5047 0.3809 0.5047 0.7104
No log 11.0 396 0.5445 0.3919 0.5445 0.7379
No log 11.0556 398 0.6054 0.4513 0.6054 0.7781
No log 11.1111 400 0.6025 0.4513 0.6025 0.7762
No log 11.1667 402 0.5949 0.4270 0.5949 0.7713
No log 11.2222 404 0.5808 0.4270 0.5808 0.7621
No log 11.2778 406 0.5662 0.3622 0.5662 0.7525
No log 11.3333 408 0.5556 0.3782 0.5556 0.7454
No log 11.3889 410 0.5749 0.3782 0.5749 0.7582
No log 11.4444 412 0.6269 0.4414 0.6269 0.7918
No log 11.5 414 0.7474 0.3359 0.7474 0.8645
No log 11.5556 416 0.7862 0.3294 0.7862 0.8867
No log 11.6111 418 0.6937 0.3770 0.6937 0.8329
No log 11.6667 420 0.5941 0.3622 0.5941 0.7708
No log 11.7222 422 0.5756 0.4086 0.5756 0.7587
No log 11.7778 424 0.6105 0.4375 0.6105 0.7813
No log 11.8333 426 0.6845 0.4835 0.6845 0.8273
No log 11.8889 428 0.7120 0.3586 0.7120 0.8438
No log 11.9444 430 0.6172 0.4375 0.6172 0.7856
No log 12.0 432 0.5504 0.4659 0.5504 0.7419
No log 12.0556 434 0.5561 0.4681 0.5561 0.7457
No log 12.1111 436 0.5520 0.4538 0.5520 0.7429
No log 12.1667 438 0.6122 0.3701 0.6122 0.7825
No log 12.2222 440 0.8007 0.3665 0.8007 0.8948
No log 12.2778 442 0.9434 0.3697 0.9434 0.9713
No log 12.3333 444 0.8034 0.3731 0.8034 0.8964
No log 12.3889 446 0.6127 0.4352 0.6127 0.7828
No log 12.4444 448 0.5406 0.4972 0.5406 0.7353
No log 12.5 450 0.5655 0.5587 0.5655 0.7520
No log 12.5556 452 0.5482 0.5493 0.5482 0.7404
No log 12.6111 454 0.5526 0.4314 0.5526 0.7434
No log 12.6667 456 0.6040 0.4774 0.6040 0.7772
No log 12.7222 458 0.6165 0.4502 0.6165 0.7852
No log 12.7778 460 0.5702 0.4524 0.5702 0.7551
No log 12.8333 462 0.5295 0.4659 0.5295 0.7277
No log 12.8889 464 0.5263 0.5305 0.5263 0.7255
No log 12.9444 466 0.5218 0.5114 0.5218 0.7223
No log 13.0 468 0.5320 0.4837 0.5320 0.7294
No log 13.0556 470 0.6141 0.4997 0.6141 0.7837
No log 13.1111 472 0.6822 0.4949 0.6822 0.8259
No log 13.1667 474 0.6493 0.5455 0.6493 0.8058
No log 13.2222 476 0.5540 0.5195 0.5540 0.7443
No log 13.2778 478 0.5218 0.5022 0.5218 0.7224
No log 13.3333 480 0.5316 0.4795 0.5316 0.7291
No log 13.3889 482 0.5295 0.4795 0.5295 0.7277
No log 13.4444 484 0.5296 0.4378 0.5296 0.7277
No log 13.5 486 0.5225 0.4591 0.5225 0.7229
No log 13.5556 488 0.5174 0.4659 0.5174 0.7193
No log 13.6111 490 0.5198 0.4378 0.5198 0.7210
No log 13.6667 492 0.5778 0.4100 0.5778 0.7602
No log 13.7222 494 0.6460 0.4911 0.6460 0.8038
No log 13.7778 496 0.6805 0.4444 0.6805 0.8250
No log 13.8333 498 0.6568 0.4684 0.6568 0.8104
0.3227 13.8889 500 0.5390 0.5195 0.5390 0.7342
0.3227 13.9444 502 0.4918 0.5422 0.4918 0.7013
0.3227 14.0 504 0.4744 0.5457 0.4744 0.6888
0.3227 14.0556 506 0.4768 0.5457 0.4768 0.6905
0.3227 14.1111 508 0.4887 0.5422 0.4887 0.6990
0.3227 14.1667 510 0.5270 0.5104 0.5270 0.7259
0.3227 14.2222 512 0.5589 0.4597 0.5589 0.7476
0.3227 14.2778 514 0.5538 0.4597 0.5538 0.7442
0.3227 14.3333 516 0.5305 0.5104 0.5305 0.7283
0.3227 14.3889 518 0.4958 0.6115 0.4958 0.7042
0.3227 14.4444 520 0.4664 0.5846 0.4664 0.6829
0.3227 14.5 522 0.4645 0.5510 0.4645 0.6815
0.3227 14.5556 524 0.4730 0.5457 0.4730 0.6878
0.3227 14.6111 526 0.5468 0.4774 0.5468 0.7395
0.3227 14.6667 528 0.6056 0.4749 0.6056 0.7782
0.3227 14.7222 530 0.5771 0.4835 0.5771 0.7597
0.3227 14.7778 532 0.5175 0.4614 0.5175 0.7194
0.3227 14.8333 534 0.4929 0.4847 0.4929 0.7021

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k14_task7_organization

Finetuned
(4023)
this model