ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k8_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5349
  • Qwk: 0.5816
  • Mse: 0.5349
  • Rmse: 0.7314

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.05 2 2.5548 -0.0262 2.5548 1.5984
No log 0.1 4 1.3626 0.0750 1.3626 1.1673
No log 0.15 6 1.0366 -0.0550 1.0366 1.0181
No log 0.2 8 0.8352 0.0025 0.8352 0.9139
No log 0.25 10 0.8067 0.1866 0.8067 0.8982
No log 0.3 12 0.6454 0.0851 0.6454 0.8033
No log 0.35 14 0.6627 0.2095 0.6627 0.8140
No log 0.4 16 0.6310 0.1754 0.6310 0.7943
No log 0.45 18 0.6069 0.2036 0.6069 0.7790
No log 0.5 20 0.7151 0.3090 0.7151 0.8456
No log 0.55 22 0.7616 0.3013 0.7616 0.8727
No log 0.6 24 0.7548 0.2942 0.7548 0.8688
No log 0.65 26 0.8476 0.4085 0.8476 0.9207
No log 0.7 28 1.0819 0.3463 1.0819 1.0401
No log 0.75 30 1.0076 0.3567 1.0076 1.0038
No log 0.8 32 0.7469 0.2736 0.7469 0.8642
No log 0.85 34 0.8427 0.2672 0.8427 0.9180
No log 0.9 36 1.1896 0.2679 1.1896 1.0907
No log 0.95 38 1.2362 0.2155 1.2362 1.1118
No log 1.0 40 0.9138 0.2113 0.9138 0.9559
No log 1.05 42 0.7771 0.2672 0.7771 0.8815
No log 1.1 44 0.7558 0.2736 0.7558 0.8694
No log 1.15 46 0.6472 0.1321 0.6472 0.8045
No log 1.2 48 0.6104 0.2024 0.6104 0.7813
No log 1.25 50 0.6148 0.1236 0.6148 0.7841
No log 1.3 52 0.6256 0.0840 0.6256 0.7910
No log 1.35 54 0.6087 0.1942 0.6087 0.7802
No log 1.4 56 0.6335 0.2353 0.6335 0.7959
No log 1.45 58 0.5960 0.3366 0.5960 0.7720
No log 1.5 60 0.6061 0.2947 0.6061 0.7785
No log 1.55 62 0.7613 0.4375 0.7613 0.8725
No log 1.6 64 0.7662 0.4321 0.7662 0.8753
No log 1.65 66 0.6230 0.4307 0.6230 0.7893
No log 1.7 68 0.5822 0.3375 0.5822 0.7630
No log 1.75 70 0.5784 0.3375 0.5784 0.7605
No log 1.8 72 0.6543 0.4161 0.6543 0.8089
No log 1.85 74 0.7042 0.4302 0.7042 0.8392
No log 1.9 76 0.6364 0.4139 0.6364 0.7977
No log 1.95 78 0.5878 0.4058 0.5878 0.7667
No log 2.0 80 0.6228 0.4436 0.6228 0.7892
No log 2.05 82 0.8433 0.4413 0.8433 0.9183
No log 2.1 84 0.9357 0.4134 0.9357 0.9673
No log 2.15 86 0.9339 0.4326 0.9339 0.9664
No log 2.2 88 0.8156 0.4217 0.8156 0.9031
No log 2.25 90 0.5405 0.6020 0.5405 0.7352
No log 2.3 92 0.5674 0.4911 0.5674 0.7533
No log 2.35 94 0.5451 0.5639 0.5451 0.7383
No log 2.4 96 0.6012 0.4355 0.6012 0.7754
No log 2.45 98 0.6733 0.3851 0.6733 0.8206
No log 2.5 100 0.5784 0.4059 0.5784 0.7605
No log 2.55 102 0.5479 0.4717 0.5479 0.7402
No log 2.6 104 0.5340 0.5127 0.5340 0.7307
No log 2.65 106 0.5554 0.5338 0.5554 0.7453
No log 2.7 108 0.5874 0.5133 0.5874 0.7664
No log 2.75 110 0.5316 0.5578 0.5316 0.7291
No log 2.8 112 0.6414 0.5003 0.6414 0.8009
No log 2.85 114 0.8940 0.4167 0.8940 0.9455
No log 2.9 116 0.8259 0.4385 0.8259 0.9088
No log 2.95 118 0.5770 0.5014 0.5770 0.7596
No log 3.0 120 0.5152 0.5319 0.5152 0.7178
No log 3.05 122 0.5626 0.4729 0.5626 0.7501
No log 3.1 124 0.7359 0.4906 0.7359 0.8579
No log 3.15 126 0.7472 0.4735 0.7472 0.8644
No log 3.2 128 0.5773 0.4058 0.5773 0.7598
No log 3.25 130 0.5210 0.5681 0.5210 0.7218
No log 3.3 132 0.5254 0.5719 0.5254 0.7248
No log 3.35 134 0.5544 0.4480 0.5544 0.7446
No log 3.4 136 0.6373 0.4410 0.6373 0.7983
No log 3.45 138 0.5697 0.4575 0.5697 0.7548
No log 3.5 140 0.5247 0.5399 0.5247 0.7244
No log 3.55 142 0.5183 0.5826 0.5183 0.7200
No log 3.6 144 0.5932 0.4854 0.5932 0.7702
No log 3.65 146 0.5467 0.5770 0.5467 0.7394
No log 3.7 148 0.5864 0.5200 0.5864 0.7657
No log 3.75 150 0.9629 0.3444 0.9629 0.9813
No log 3.8 152 0.9867 0.3444 0.9867 0.9933
No log 3.85 154 0.7020 0.3827 0.7020 0.8379
No log 3.9 156 0.5420 0.4903 0.5420 0.7362
No log 3.95 158 0.5578 0.5134 0.5578 0.7469
No log 4.0 160 0.5591 0.5190 0.5591 0.7477
No log 4.05 162 0.5692 0.5136 0.5692 0.7545
No log 4.1 164 0.5259 0.4953 0.5259 0.7252
No log 4.15 166 0.5746 0.4444 0.5746 0.7580
No log 4.2 168 0.5254 0.5177 0.5254 0.7249
No log 4.25 170 0.5314 0.5319 0.5314 0.7290
No log 4.3 172 0.7120 0.4287 0.7120 0.8438
No log 4.35 174 0.7785 0.4269 0.7785 0.8823
No log 4.4 176 0.5861 0.5228 0.5861 0.7656
No log 4.45 178 0.5632 0.5616 0.5632 0.7505
No log 4.5 180 0.6248 0.5616 0.6248 0.7904
No log 4.55 182 0.7461 0.4589 0.7461 0.8637
No log 4.6 184 0.7167 0.4409 0.7167 0.8466
No log 4.65 186 0.5761 0.5319 0.5761 0.7590
No log 4.7 188 0.5766 0.4664 0.5766 0.7593
No log 4.75 190 0.6054 0.4502 0.6054 0.7781
No log 4.8 192 0.5230 0.4020 0.5230 0.7232
No log 4.85 194 0.5424 0.5368 0.5424 0.7365
No log 4.9 196 0.6352 0.4537 0.6352 0.7970
No log 4.95 198 0.6318 0.4747 0.6318 0.7949
No log 5.0 200 0.5695 0.4795 0.5695 0.7546
No log 5.05 202 0.5634 0.5357 0.5634 0.7506
No log 5.1 204 0.5723 0.5140 0.5723 0.7565
No log 5.15 206 0.5488 0.4265 0.5488 0.7408
No log 5.2 208 0.5824 0.4802 0.5824 0.7632
No log 5.25 210 0.5619 0.4849 0.5619 0.7496
No log 5.3 212 0.5441 0.3861 0.5441 0.7376
No log 5.35 214 0.5528 0.4951 0.5528 0.7435
No log 5.4 216 0.5430 0.5158 0.5430 0.7369
No log 5.45 218 0.5407 0.5121 0.5407 0.7353
No log 5.5 220 0.5350 0.5063 0.5350 0.7314
No log 5.55 222 0.5431 0.5399 0.5431 0.7370
No log 5.6 224 0.6304 0.5630 0.6304 0.7940
No log 5.65 226 0.8074 0.4397 0.8074 0.8985
No log 5.7 228 0.8804 0.3735 0.8804 0.9383
No log 5.75 230 0.7535 0.4651 0.7535 0.8681
No log 5.8 232 0.6395 0.5184 0.6395 0.7997
No log 5.85 234 0.6155 0.4807 0.6155 0.7846
No log 5.9 236 0.5736 0.4575 0.5736 0.7573
No log 5.95 238 0.5908 0.4518 0.5908 0.7686
No log 6.0 240 0.5480 0.4555 0.5480 0.7402
No log 6.05 242 0.5460 0.5158 0.5460 0.7389
No log 6.1 244 0.5549 0.5679 0.5549 0.7449
No log 6.15 246 0.5779 0.5117 0.5779 0.7602
No log 6.2 248 0.5249 0.5614 0.5249 0.7245
No log 6.25 250 0.4990 0.6027 0.4990 0.7064
No log 6.3 252 0.4879 0.6105 0.4879 0.6985
No log 6.35 254 0.4909 0.5840 0.4909 0.7006
No log 6.4 256 0.4890 0.5288 0.4890 0.6993
No log 6.45 258 0.5767 0.4837 0.5767 0.7594
No log 6.5 260 0.5311 0.4864 0.5311 0.7288
No log 6.55 262 0.4929 0.6566 0.4929 0.7020
No log 6.6 264 0.5601 0.5650 0.5601 0.7484
No log 6.65 266 0.5444 0.5031 0.5444 0.7379
No log 6.7 268 0.4913 0.6111 0.4913 0.7009
No log 6.75 270 0.5656 0.4836 0.5656 0.7521
No log 6.8 272 0.5808 0.4836 0.5808 0.7621
No log 6.85 274 0.5041 0.5411 0.5041 0.7100
No log 6.9 276 0.5163 0.5592 0.5163 0.7185
No log 6.95 278 0.5581 0.5486 0.5581 0.7471
No log 7.0 280 0.5036 0.5781 0.5036 0.7096
No log 7.05 282 0.4950 0.5488 0.4950 0.7035
No log 7.1 284 0.4990 0.5322 0.4990 0.7064
No log 7.15 286 0.4923 0.6105 0.4923 0.7017
No log 7.2 288 0.5254 0.5607 0.5254 0.7248
No log 7.25 290 0.5436 0.5471 0.5436 0.7373
No log 7.3 292 0.5230 0.6018 0.5230 0.7232
No log 7.35 294 0.5410 0.5656 0.5410 0.7355
No log 7.4 296 0.5103 0.6210 0.5103 0.7144
No log 7.45 298 0.5218 0.5345 0.5218 0.7224
No log 7.5 300 0.5339 0.5177 0.5339 0.7307
No log 7.55 302 0.5178 0.5520 0.5178 0.7196
No log 7.6 304 0.5256 0.5271 0.5256 0.7250
No log 7.65 306 0.5200 0.5342 0.5200 0.7211
No log 7.7 308 0.5051 0.6282 0.5051 0.7107
No log 7.75 310 0.5170 0.6526 0.5170 0.7191
No log 7.8 312 0.5435 0.6315 0.5435 0.7372
No log 7.85 314 0.5143 0.6876 0.5143 0.7171
No log 7.9 316 0.5007 0.6706 0.5007 0.7076
No log 7.95 318 0.5017 0.6387 0.5017 0.7083
No log 8.0 320 0.4896 0.6662 0.4896 0.6997
No log 8.05 322 0.5333 0.5985 0.5333 0.7303
No log 8.1 324 0.5503 0.6096 0.5503 0.7418
No log 8.15 326 0.6250 0.5323 0.6250 0.7906
No log 8.2 328 0.6994 0.5232 0.6994 0.8363
No log 8.25 330 0.6239 0.5938 0.6239 0.7899
No log 8.3 332 0.5947 0.6543 0.5947 0.7712
No log 8.35 334 0.5698 0.6404 0.5698 0.7548
No log 8.4 336 0.5768 0.6544 0.5768 0.7595
No log 8.45 338 0.7005 0.4918 0.7005 0.8369
No log 8.5 340 0.8094 0.4413 0.8094 0.8997
No log 8.55 342 0.7457 0.4413 0.7457 0.8635
No log 8.6 344 0.5803 0.5428 0.5803 0.7618
No log 8.65 346 0.4912 0.6495 0.4912 0.7009
No log 8.7 348 0.4860 0.5612 0.4860 0.6971
No log 8.75 350 0.4889 0.6186 0.4889 0.6992
No log 8.8 352 0.5000 0.6503 0.5000 0.7071
No log 8.85 354 0.5278 0.6096 0.5278 0.7265
No log 8.9 356 0.5417 0.6096 0.5417 0.7360
No log 8.95 358 0.5463 0.5861 0.5463 0.7391
No log 9.0 360 0.5739 0.5355 0.5739 0.7576
No log 9.05 362 0.5620 0.5355 0.5620 0.7496
No log 9.1 364 0.5269 0.6182 0.5269 0.7259
No log 9.15 366 0.5085 0.5795 0.5085 0.7131
No log 9.2 368 0.4875 0.6491 0.4875 0.6982
No log 9.25 370 0.4621 0.5951 0.4621 0.6798
No log 9.3 372 0.4593 0.6032 0.4593 0.6777
No log 9.35 374 0.4582 0.6039 0.4582 0.6769
No log 9.4 376 0.5062 0.5592 0.5062 0.7115
No log 9.45 378 0.5098 0.5252 0.5098 0.7140
No log 9.5 380 0.4785 0.6118 0.4785 0.6917
No log 9.55 382 0.4641 0.6078 0.4641 0.6812
No log 9.6 384 0.4650 0.5923 0.4650 0.6819
No log 9.65 386 0.4650 0.5923 0.4650 0.6819
No log 9.7 388 0.4810 0.5544 0.4810 0.6936
No log 9.75 390 0.4628 0.5836 0.4628 0.6803
No log 9.8 392 0.4832 0.5414 0.4832 0.6951
No log 9.85 394 0.5530 0.5384 0.5530 0.7436
No log 9.9 396 0.5250 0.5584 0.5250 0.7246
No log 9.95 398 0.4799 0.6210 0.4799 0.6928
No log 10.0 400 0.4519 0.6467 0.4519 0.6722
No log 10.05 402 0.4571 0.6383 0.4571 0.6761
No log 10.1 404 0.4717 0.6183 0.4717 0.6868
No log 10.15 406 0.4901 0.5918 0.4901 0.7001
No log 10.2 408 0.4710 0.6170 0.4710 0.6863
No log 10.25 410 0.4859 0.5741 0.4859 0.6971
No log 10.3 412 0.5207 0.6109 0.5207 0.7216
No log 10.35 414 0.4919 0.5836 0.4919 0.7014
No log 10.4 416 0.4769 0.5665 0.4769 0.6906
No log 10.45 418 0.4809 0.5665 0.4809 0.6935
No log 10.5 420 0.4800 0.5899 0.4800 0.6928
No log 10.55 422 0.4785 0.5899 0.4785 0.6917
No log 10.6 424 0.4771 0.5926 0.4771 0.6907
No log 10.65 426 0.4951 0.6300 0.4951 0.7037
No log 10.7 428 0.4941 0.6300 0.4941 0.7029
No log 10.75 430 0.4816 0.5600 0.4816 0.6940
No log 10.8 432 0.5396 0.5144 0.5396 0.7345
No log 10.85 434 0.5451 0.5794 0.5451 0.7383
No log 10.9 436 0.5054 0.6295 0.5054 0.7109
No log 10.95 438 0.4758 0.5860 0.4758 0.6898
No log 11.0 440 0.4827 0.6452 0.4827 0.6948
No log 11.05 442 0.5019 0.5920 0.5019 0.7084
No log 11.1 444 0.4898 0.6282 0.4898 0.6999
No log 11.15 446 0.4810 0.6118 0.4810 0.6935
No log 11.2 448 0.5116 0.6079 0.5116 0.7152
No log 11.25 450 0.4964 0.6419 0.4964 0.7046
No log 11.3 452 0.4894 0.6419 0.4894 0.6996
No log 11.35 454 0.4699 0.6407 0.4699 0.6855
No log 11.4 456 0.4535 0.6655 0.4535 0.6734
No log 11.45 458 0.4505 0.6377 0.4505 0.6712
No log 11.5 460 0.4457 0.6655 0.4457 0.6676
No log 11.55 462 0.4497 0.6727 0.4497 0.6706
No log 11.6 464 0.4674 0.6706 0.4674 0.6837
No log 11.65 466 0.4877 0.6549 0.4877 0.6983
No log 11.7 468 0.4801 0.7059 0.4801 0.6929
No log 11.75 470 0.5364 0.6390 0.5364 0.7324
No log 11.8 472 0.5792 0.5710 0.5792 0.7611
No log 11.85 474 0.5192 0.6361 0.5192 0.7206
No log 11.9 476 0.4716 0.6339 0.4716 0.6867
No log 11.95 478 0.5019 0.5076 0.5019 0.7084
No log 12.0 480 0.4877 0.5708 0.4877 0.6984
No log 12.05 482 0.4797 0.6757 0.4797 0.6926
No log 12.1 484 0.4860 0.6032 0.4860 0.6972
No log 12.15 486 0.4799 0.6757 0.4799 0.6927
No log 12.2 488 0.4754 0.6543 0.4754 0.6895
No log 12.25 490 0.4678 0.6648 0.4678 0.6840
No log 12.3 492 0.4629 0.6554 0.4629 0.6803
No log 12.35 494 0.4761 0.6402 0.4761 0.6900
No log 12.4 496 0.4745 0.6402 0.4745 0.6889
No log 12.45 498 0.4588 0.6771 0.4588 0.6773
0.2988 12.5 500 0.4536 0.6377 0.4536 0.6735
0.2988 12.55 502 0.4572 0.6563 0.4572 0.6762
0.2988 12.6 504 0.4722 0.6503 0.4722 0.6871
0.2988 12.65 506 0.5174 0.6206 0.5174 0.7193
0.2988 12.7 508 0.5039 0.6283 0.5039 0.7099
0.2988 12.75 510 0.4624 0.6389 0.4624 0.6800
0.2988 12.8 512 0.4510 0.6747 0.4510 0.6715
0.2988 12.85 514 0.4669 0.5386 0.4669 0.6833
0.2988 12.9 516 0.4643 0.5386 0.4643 0.6814
0.2988 12.95 518 0.4578 0.6661 0.4578 0.6766
0.2988 13.0 520 0.4772 0.5999 0.4772 0.6908
0.2988 13.05 522 0.4791 0.5999 0.4791 0.6922
0.2988 13.1 524 0.4818 0.6395 0.4818 0.6941
0.2988 13.15 526 0.4828 0.6914 0.4828 0.6948
0.2988 13.2 528 0.4792 0.7003 0.4792 0.6922
0.2988 13.25 530 0.4760 0.6828 0.4760 0.6900
0.2988 13.3 532 0.4717 0.6555 0.4717 0.6868
0.2988 13.35 534 0.4702 0.6426 0.4702 0.6857
0.2988 13.4 536 0.4875 0.6087 0.4875 0.6982
0.2988 13.45 538 0.4722 0.5671 0.4722 0.6871
0.2988 13.5 540 0.4598 0.6747 0.4598 0.6781
0.2988 13.55 542 0.4757 0.5970 0.4757 0.6897
0.2988 13.6 544 0.5241 0.5895 0.5241 0.7239
0.2988 13.65 546 0.5018 0.5970 0.5018 0.7084
0.2988 13.7 548 0.4636 0.6643 0.4636 0.6809
0.2988 13.75 550 0.4730 0.6643 0.4730 0.6878
0.2988 13.8 552 0.4747 0.6643 0.4747 0.6890
0.2988 13.85 554 0.4816 0.6158 0.4816 0.6940
0.2988 13.9 556 0.4716 0.6467 0.4716 0.6868
0.2988 13.95 558 0.4751 0.6566 0.4751 0.6893
0.2988 14.0 560 0.4752 0.6566 0.4752 0.6893
0.2988 14.05 562 0.4813 0.6467 0.4813 0.6937
0.2988 14.1 564 0.5096 0.5770 0.5096 0.7138
0.2988 14.15 566 0.5845 0.5368 0.5845 0.7645
0.2988 14.2 568 0.5822 0.5498 0.5822 0.7630
0.2988 14.25 570 0.5775 0.4951 0.5775 0.7600
0.2988 14.3 572 0.5349 0.5816 0.5349 0.7314

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k8_task7_organization

Finetuned
(4019)
this model