ArabicNewSplits6_FineTuningAraBERT_run3_AugV5_k12_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6501
  • Qwk: 0.4269
  • Mse: 0.6501
  • Rmse: 0.8063

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0328 2 4.5181 -0.0269 4.5181 2.1256
No log 0.0656 4 2.7509 0.0341 2.7509 1.6586
No log 0.0984 6 1.5624 -0.0218 1.5624 1.2500
No log 0.1311 8 1.2150 0.0190 1.2150 1.1022
No log 0.1639 10 1.0114 -0.0111 1.0114 1.0057
No log 0.1967 12 1.1651 0.0045 1.1651 1.0794
No log 0.2295 14 1.4342 0.0410 1.4342 1.1976
No log 0.2623 16 1.1411 0.0307 1.1411 1.0682
No log 0.2951 18 0.9116 -0.0037 0.9116 0.9548
No log 0.3279 20 0.7538 0.1764 0.7538 0.8682
No log 0.3607 22 0.7194 0.1778 0.7194 0.8482
No log 0.3934 24 0.7055 0.2376 0.7055 0.8399
No log 0.4262 26 0.6691 0.2853 0.6691 0.8180
No log 0.4590 28 0.6820 0.2209 0.6820 0.8258
No log 0.4918 30 0.8517 0.0975 0.8517 0.9229
No log 0.5246 32 1.3858 0.0166 1.3858 1.1772
No log 0.5574 34 1.6021 0.0602 1.6021 1.2657
No log 0.5902 36 1.3132 0.0361 1.3132 1.1460
No log 0.6230 38 0.9284 0.0 0.9284 0.9636
No log 0.6557 40 0.7327 0.2413 0.7327 0.8560
No log 0.6885 42 0.6700 0.3663 0.6700 0.8185
No log 0.7213 44 0.6598 0.375 0.6598 0.8123
No log 0.7541 46 0.6434 0.3577 0.6434 0.8021
No log 0.7869 48 0.6948 0.2775 0.6948 0.8336
No log 0.8197 50 0.8021 0.2255 0.8021 0.8956
No log 0.8525 52 0.8821 0.2093 0.8821 0.9392
No log 0.8852 54 0.8227 0.2267 0.8227 0.9070
No log 0.9180 56 0.6641 0.3406 0.6641 0.8149
No log 0.9508 58 0.6747 0.2961 0.6747 0.8214
No log 0.9836 60 0.8230 0.1468 0.8230 0.9072
No log 1.0164 62 1.0372 0.1554 1.0372 1.0184
No log 1.0492 64 0.9897 0.1755 0.9897 0.9948
No log 1.0820 66 0.8836 0.1811 0.8836 0.9400
No log 1.1148 68 0.6633 0.4004 0.6633 0.8144
No log 1.1475 70 0.6456 0.3957 0.6456 0.8035
No log 1.1803 72 0.6375 0.4651 0.6375 0.7985
No log 1.2131 74 0.6468 0.4743 0.6468 0.8042
No log 1.2459 76 0.6455 0.3963 0.6455 0.8034
No log 1.2787 78 0.6502 0.3875 0.6502 0.8063
No log 1.3115 80 0.6423 0.3846 0.6423 0.8014
No log 1.3443 82 0.6330 0.4769 0.6330 0.7956
No log 1.3770 84 0.7123 0.3559 0.7123 0.8440
No log 1.4098 86 0.7679 0.2997 0.7679 0.8763
No log 1.4426 88 0.6686 0.3616 0.6686 0.8177
No log 1.4754 90 0.5635 0.5313 0.5635 0.7507
No log 1.5082 92 0.5768 0.5638 0.5768 0.7594
No log 1.5410 94 0.6072 0.4474 0.6072 0.7792
No log 1.5738 96 0.6033 0.4838 0.6033 0.7767
No log 1.6066 98 0.6252 0.4626 0.6252 0.7907
No log 1.6393 100 0.6400 0.4776 0.6400 0.8000
No log 1.6721 102 0.6702 0.4623 0.6702 0.8186
No log 1.7049 104 0.7245 0.4762 0.7245 0.8511
No log 1.7377 106 0.7617 0.5022 0.7617 0.8728
No log 1.7705 108 0.7806 0.4727 0.7806 0.8835
No log 1.8033 110 0.7646 0.4670 0.7646 0.8744
No log 1.8361 112 0.6878 0.4799 0.6878 0.8293
No log 1.8689 114 0.6473 0.4762 0.6473 0.8046
No log 1.9016 116 0.6492 0.4227 0.6492 0.8057
No log 1.9344 118 0.6478 0.4557 0.6478 0.8049
No log 1.9672 120 0.6223 0.4714 0.6223 0.7889
No log 2.0 122 0.6079 0.4516 0.6079 0.7797
No log 2.0328 124 0.6212 0.4674 0.6212 0.7882
No log 2.0656 126 0.6164 0.4861 0.6164 0.7851
No log 2.0984 128 0.5911 0.4532 0.5911 0.7688
No log 2.1311 130 0.6511 0.4064 0.6511 0.8069
No log 2.1639 132 0.6986 0.3976 0.6986 0.8358
No log 2.1967 134 0.6963 0.4200 0.6963 0.8345
No log 2.2295 136 0.7438 0.3851 0.7438 0.8624
No log 2.2623 138 0.7748 0.3924 0.7748 0.8802
No log 2.2951 140 0.7555 0.4307 0.7555 0.8692
No log 2.3279 142 0.6359 0.4364 0.6359 0.7975
No log 2.3607 144 0.6004 0.5171 0.6004 0.7748
No log 2.3934 146 0.6986 0.4947 0.6986 0.8358
No log 2.4262 148 0.8572 0.4240 0.8572 0.9259
No log 2.4590 150 0.8762 0.4047 0.8762 0.9361
No log 2.4918 152 0.8143 0.4172 0.8143 0.9024
No log 2.5246 154 0.7836 0.4621 0.7836 0.8852
No log 2.5574 156 0.7067 0.5188 0.7067 0.8406
No log 2.5902 158 0.6691 0.4793 0.6691 0.8180
No log 2.6230 160 0.6817 0.4683 0.6817 0.8256
No log 2.6557 162 0.7419 0.4349 0.7419 0.8613
No log 2.6885 164 0.8170 0.4264 0.8170 0.9039
No log 2.7213 166 0.8228 0.4183 0.8228 0.9071
No log 2.7541 168 0.7748 0.4143 0.7748 0.8802
No log 2.7869 170 0.6996 0.4549 0.6996 0.8364
No log 2.8197 172 0.7060 0.4348 0.7060 0.8402
No log 2.8525 174 0.7255 0.4537 0.7255 0.8518
No log 2.8852 176 0.6564 0.4470 0.6564 0.8102
No log 2.9180 178 0.6242 0.4795 0.6242 0.7901
No log 2.9508 180 0.6424 0.4291 0.6424 0.8015
No log 2.9836 182 0.6705 0.4199 0.6705 0.8188
No log 3.0164 184 0.6555 0.4198 0.6555 0.8096
No log 3.0492 186 0.6117 0.4814 0.6117 0.7821
No log 3.0820 188 0.6132 0.4670 0.6132 0.7831
No log 3.1148 190 0.6282 0.4408 0.6282 0.7926
No log 3.1475 192 0.6589 0.4449 0.6589 0.8117
No log 3.1803 194 0.7026 0.4420 0.7026 0.8382
No log 3.2131 196 0.7265 0.4806 0.7265 0.8524
No log 3.2459 198 0.7358 0.4703 0.7358 0.8578
No log 3.2787 200 0.7253 0.4641 0.7253 0.8517
No log 3.3115 202 0.6949 0.4549 0.6949 0.8336
No log 3.3443 204 0.6605 0.4362 0.6605 0.8127
No log 3.3770 206 0.6565 0.4814 0.6565 0.8102
No log 3.4098 208 0.7278 0.4286 0.7278 0.8531
No log 3.4426 210 0.7861 0.4013 0.7861 0.8866
No log 3.4754 212 0.7654 0.4243 0.7654 0.8749
No log 3.5082 214 0.6704 0.4311 0.6704 0.8188
No log 3.5410 216 0.6231 0.4348 0.6231 0.7894
No log 3.5738 218 0.6697 0.4191 0.6697 0.8183
No log 3.6066 220 0.6995 0.4553 0.6995 0.8363
No log 3.6393 222 0.6465 0.4539 0.6465 0.8040
No log 3.6721 224 0.6540 0.4725 0.6540 0.8087
No log 3.7049 226 0.7105 0.4457 0.7105 0.8429
No log 3.7377 228 0.6916 0.4288 0.6916 0.8316
No log 3.7705 230 0.6805 0.4398 0.6805 0.8249
No log 3.8033 232 0.7065 0.4606 0.7065 0.8405
No log 3.8361 234 0.7127 0.4685 0.7127 0.8442
No log 3.8689 236 0.6691 0.4790 0.6691 0.8180
No log 3.9016 238 0.5969 0.4758 0.5969 0.7726
No log 3.9344 240 0.5709 0.4843 0.5709 0.7556
No log 3.9672 242 0.5969 0.4085 0.5969 0.7726
No log 4.0 244 0.6034 0.4743 0.6034 0.7768
No log 4.0328 246 0.6358 0.4997 0.6358 0.7974
No log 4.0656 248 0.7418 0.4950 0.7418 0.8613
No log 4.0984 250 0.8468 0.4758 0.8468 0.9202
No log 4.1311 252 0.8470 0.4903 0.8470 0.9203
No log 4.1639 254 0.8071 0.4983 0.8071 0.8984
No log 4.1967 256 0.7257 0.5142 0.7257 0.8519
No log 4.2295 258 0.6647 0.4912 0.6647 0.8153
No log 4.2623 260 0.6624 0.5171 0.6624 0.8139
No log 4.2951 262 0.6688 0.5092 0.6688 0.8178
No log 4.3279 264 0.7235 0.4906 0.7235 0.8506
No log 4.3607 266 0.7808 0.4700 0.7808 0.8836
No log 4.3934 268 0.7811 0.4186 0.7811 0.8838
No log 4.4262 270 0.7022 0.4683 0.7022 0.8380
No log 4.4590 272 0.6357 0.4769 0.6357 0.7973
No log 4.4918 274 0.6237 0.4527 0.6237 0.7898
No log 4.5246 276 0.6344 0.4843 0.6344 0.7965
No log 4.5574 278 0.6676 0.4864 0.6676 0.8171
No log 4.5902 280 0.7532 0.4517 0.7532 0.8679
No log 4.6230 282 0.8594 0.4243 0.8594 0.9270
No log 4.6557 284 0.8915 0.4223 0.8915 0.9442
No log 4.6885 286 0.8627 0.4243 0.8627 0.9288
No log 4.7213 288 0.8465 0.4243 0.8465 0.9201
No log 4.7541 290 0.7619 0.4776 0.7619 0.8729
No log 4.7869 292 0.6764 0.4707 0.6764 0.8225
No log 4.8197 294 0.6486 0.4704 0.6486 0.8053
No log 4.8525 296 0.6441 0.4984 0.6441 0.8026
No log 4.8852 298 0.6447 0.5074 0.6447 0.8029
No log 4.9180 300 0.6604 0.4335 0.6604 0.8126
No log 4.9508 302 0.6725 0.4120 0.6725 0.8201
No log 4.9836 304 0.6495 0.4335 0.6495 0.8059
No log 5.0164 306 0.6351 0.4148 0.6351 0.7969
No log 5.0492 308 0.6350 0.4338 0.6350 0.7969
No log 5.0820 310 0.6534 0.4520 0.6534 0.8084
No log 5.1148 312 0.6773 0.4557 0.6773 0.8230
No log 5.1475 314 0.7029 0.4313 0.7029 0.8384
No log 5.1803 316 0.7306 0.4433 0.7306 0.8547
No log 5.2131 318 0.7701 0.4704 0.7701 0.8775
No log 5.2459 320 0.8117 0.4530 0.8117 0.9009
No log 5.2787 322 0.8318 0.4713 0.8318 0.9120
No log 5.3115 324 0.7983 0.4719 0.7983 0.8935
No log 5.3443 326 0.7593 0.4730 0.7593 0.8714
No log 5.3770 328 0.7558 0.4824 0.7558 0.8693
No log 5.4098 330 0.7225 0.4911 0.7225 0.8500
No log 5.4426 332 0.7059 0.4938 0.7059 0.8402
No log 5.4754 334 0.6702 0.4759 0.6702 0.8187
No log 5.5082 336 0.6352 0.4822 0.6352 0.7970
No log 5.5410 338 0.6310 0.4822 0.6310 0.7943
No log 5.5738 340 0.6162 0.4361 0.6162 0.7850
No log 5.6066 342 0.6054 0.4723 0.6054 0.7781
No log 5.6393 344 0.6150 0.4716 0.6150 0.7842
No log 5.6721 346 0.6493 0.4335 0.6493 0.8058
No log 5.7049 348 0.6689 0.4487 0.6689 0.8179
No log 5.7377 350 0.6761 0.4269 0.6761 0.8223
No log 5.7705 352 0.6651 0.3923 0.6651 0.8156
No log 5.8033 354 0.6535 0.4810 0.6535 0.8084
No log 5.8361 356 0.6604 0.4187 0.6604 0.8127
No log 5.8689 358 0.6931 0.4183 0.6931 0.8325
No log 5.9016 360 0.7280 0.4268 0.7280 0.8532
No log 5.9344 362 0.7410 0.4268 0.7410 0.8608
No log 5.9672 364 0.7232 0.4268 0.7232 0.8504
No log 6.0 366 0.7223 0.4757 0.7223 0.8499
No log 6.0328 368 0.7016 0.4635 0.7016 0.8376
No log 6.0656 370 0.6799 0.4669 0.6799 0.8245
No log 6.0984 372 0.6646 0.4565 0.6646 0.8152
No log 6.1311 374 0.6591 0.4940 0.6591 0.8118
No log 6.1639 376 0.6590 0.4940 0.6590 0.8118
No log 6.1967 378 0.6611 0.4565 0.6611 0.8131
No log 6.2295 380 0.6795 0.4698 0.6795 0.8243
No log 6.2623 382 0.7325 0.4760 0.7325 0.8559
No log 6.2951 384 0.7940 0.4099 0.7940 0.8911
No log 6.3279 386 0.7995 0.4158 0.7995 0.8941
No log 6.3607 388 0.7497 0.4414 0.7497 0.8658
No log 6.3934 390 0.6932 0.4400 0.6932 0.8326
No log 6.4262 392 0.6710 0.4449 0.6710 0.8192
No log 6.4590 394 0.6732 0.4449 0.6732 0.8205
No log 6.4918 396 0.6935 0.4553 0.6935 0.8328
No log 6.5246 398 0.7237 0.4329 0.7237 0.8507
No log 6.5574 400 0.7846 0.3891 0.7846 0.8858
No log 6.5902 402 0.8146 0.3900 0.8146 0.9025
No log 6.6230 404 0.7798 0.3891 0.7798 0.8831
No log 6.6557 406 0.7204 0.4653 0.7204 0.8488
No log 6.6885 408 0.6851 0.4246 0.6851 0.8277
No log 6.7213 410 0.6729 0.4360 0.6729 0.8203
No log 6.7541 412 0.6694 0.4448 0.6694 0.8182
No log 6.7869 414 0.6670 0.4535 0.6670 0.8167
No log 6.8197 416 0.6673 0.4099 0.6673 0.8169
No log 6.8525 418 0.6686 0.4121 0.6686 0.8177
No log 6.8852 420 0.6701 0.4342 0.6701 0.8186
No log 6.9180 422 0.6699 0.4621 0.6699 0.8185
No log 6.9508 424 0.6772 0.4337 0.6772 0.8229
No log 6.9836 426 0.6851 0.4180 0.6851 0.8277
No log 7.0164 428 0.6823 0.4246 0.6823 0.8260
No log 7.0492 430 0.6882 0.4246 0.6882 0.8296
No log 7.0820 432 0.6944 0.4465 0.6944 0.8333
No log 7.1148 434 0.6813 0.4246 0.6813 0.8254
No log 7.1475 436 0.6592 0.4578 0.6592 0.8119
No log 7.1803 438 0.6503 0.4320 0.6503 0.8064
No log 7.2131 440 0.6491 0.4409 0.6491 0.8057
No log 7.2459 442 0.6521 0.4825 0.6521 0.8075
No log 7.2787 444 0.6569 0.4315 0.6569 0.8105
No log 7.3115 446 0.6883 0.4332 0.6883 0.8296
No log 7.3443 448 0.7172 0.4740 0.7172 0.8469
No log 7.3770 450 0.7279 0.4740 0.7279 0.8532
No log 7.4098 452 0.7087 0.4473 0.7087 0.8418
No log 7.4426 454 0.6819 0.4354 0.6819 0.8258
No log 7.4754 456 0.6654 0.4466 0.6654 0.8157
No log 7.5082 458 0.6444 0.4494 0.6444 0.8027
No log 7.5410 460 0.6306 0.5010 0.6306 0.7941
No log 7.5738 462 0.6268 0.4942 0.6268 0.7917
No log 7.6066 464 0.6299 0.4942 0.6299 0.7937
No log 7.6393 466 0.6403 0.4494 0.6403 0.8002
No log 7.6721 468 0.6591 0.4383 0.6591 0.8118
No log 7.7049 470 0.6768 0.4524 0.6768 0.8227
No log 7.7377 472 0.6869 0.4499 0.6869 0.8288
No log 7.7705 474 0.6717 0.4447 0.6717 0.8196
No log 7.8033 476 0.6551 0.4383 0.6551 0.8094
No log 7.8361 478 0.6409 0.4383 0.6409 0.8006
No log 7.8689 480 0.6366 0.4633 0.6366 0.7979
No log 7.9016 482 0.6396 0.4543 0.6396 0.7997
No log 7.9344 484 0.6459 0.4383 0.6459 0.8037
No log 7.9672 486 0.6613 0.4293 0.6613 0.8132
No log 8.0 488 0.6785 0.4422 0.6785 0.8237
No log 8.0328 490 0.7012 0.4478 0.7012 0.8374
No log 8.0656 492 0.7169 0.4162 0.7169 0.8467
No log 8.0984 494 0.7127 0.4457 0.7127 0.8442
No log 8.1311 496 0.6955 0.4415 0.6955 0.8340
No log 8.1639 498 0.6887 0.4422 0.6887 0.8299
0.4415 8.1967 500 0.6787 0.4444 0.6787 0.8238
0.4415 8.2295 502 0.6616 0.4628 0.6616 0.8134
0.4415 8.2623 504 0.6539 0.4907 0.6539 0.8086
0.4415 8.2951 506 0.6535 0.4990 0.6535 0.8084
0.4415 8.3279 508 0.6590 0.4946 0.6590 0.8118
0.4415 8.3607 510 0.6631 0.4946 0.6631 0.8143
0.4415 8.3934 512 0.6640 0.4946 0.6640 0.8149
0.4415 8.4262 514 0.6656 0.4766 0.6656 0.8159
0.4415 8.4590 516 0.6670 0.4591 0.6670 0.8167
0.4415 8.4918 518 0.6638 0.4675 0.6638 0.8148
0.4415 8.5246 520 0.6620 0.4591 0.6620 0.8137
0.4415 8.5574 522 0.6571 0.4574 0.6571 0.8106
0.4415 8.5902 524 0.6533 0.4557 0.6533 0.8082
0.4415 8.6230 526 0.6575 0.4204 0.6575 0.8109
0.4415 8.6557 528 0.6611 0.4114 0.6611 0.8131
0.4415 8.6885 530 0.6635 0.4181 0.6635 0.8145
0.4415 8.7213 532 0.6611 0.4114 0.6611 0.8131
0.4415 8.7541 534 0.6611 0.4269 0.6611 0.8131
0.4415 8.7869 536 0.6620 0.4269 0.6620 0.8136
0.4415 8.8197 538 0.6591 0.4401 0.6591 0.8119
0.4415 8.8525 540 0.6550 0.4570 0.6550 0.8093
0.4415 8.8852 542 0.6522 0.4570 0.6522 0.8076
0.4415 8.9180 544 0.6552 0.4570 0.6552 0.8095
0.4415 8.9508 546 0.6559 0.4401 0.6559 0.8099
0.4415 8.9836 548 0.6562 0.4379 0.6562 0.8101
0.4415 9.0164 550 0.6568 0.4379 0.6568 0.8104
0.4415 9.0492 552 0.6570 0.4379 0.6570 0.8105
0.4415 9.0820 554 0.6577 0.4379 0.6577 0.8110
0.4415 9.1148 556 0.6618 0.4269 0.6618 0.8135
0.4415 9.1475 558 0.6675 0.4422 0.6675 0.8170
0.4415 9.1803 560 0.6710 0.4334 0.6710 0.8191
0.4415 9.2131 562 0.6740 0.4334 0.6740 0.8210
0.4415 9.2459 564 0.6788 0.4333 0.6788 0.8239
0.4415 9.2787 566 0.6791 0.4333 0.6791 0.8241
0.4415 9.3115 568 0.6734 0.4333 0.6734 0.8206
0.4415 9.3443 570 0.6693 0.4334 0.6693 0.8181
0.4415 9.3770 572 0.6626 0.4334 0.6626 0.8140
0.4415 9.4098 574 0.6537 0.4422 0.6537 0.8085
0.4415 9.4426 576 0.6452 0.4269 0.6452 0.8033
0.4415 9.4754 578 0.6396 0.4358 0.6396 0.7998
0.4415 9.5082 580 0.6360 0.4535 0.6360 0.7975
0.4415 9.5410 582 0.6336 0.4628 0.6336 0.7960
0.4415 9.5738 584 0.6321 0.4561 0.6321 0.7950
0.4415 9.6066 586 0.6320 0.4561 0.6320 0.7950
0.4415 9.6393 588 0.6336 0.4561 0.6336 0.7960
0.4415 9.6721 590 0.6363 0.4628 0.6363 0.7977
0.4415 9.7049 592 0.6396 0.4379 0.6396 0.7998
0.4415 9.7377 594 0.6416 0.4379 0.6416 0.8010
0.4415 9.7705 596 0.6433 0.4379 0.6433 0.8021
0.4415 9.8033 598 0.6452 0.4379 0.6452 0.8032
0.4415 9.8361 600 0.6465 0.4358 0.6465 0.8041
0.4415 9.8689 602 0.6478 0.4358 0.6478 0.8049
0.4415 9.9016 604 0.6487 0.4269 0.6487 0.8054
0.4415 9.9344 606 0.6494 0.4269 0.6494 0.8059
0.4415 9.9672 608 0.6499 0.4269 0.6499 0.8062
0.4415 10.0 610 0.6501 0.4269 0.6501 0.8063

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run3_AugV5_k12_task2_organization

Finetuned
(4023)
this model