ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k12_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4657
  • Qwk: 0.5003
  • Mse: 0.4657
  • Rmse: 0.6824

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0333 2 2.4964 -0.0297 2.4964 1.5800
No log 0.0667 4 1.2875 0.1719 1.2875 1.1347
No log 0.1 6 0.6737 0.2142 0.6737 0.8208
No log 0.1333 8 0.7118 0.1541 0.7118 0.8437
No log 0.1667 10 0.7477 0.1790 0.7477 0.8647
No log 0.2 12 0.7286 0.1498 0.7286 0.8536
No log 0.2333 14 0.7299 0.0937 0.7299 0.8543
No log 0.2667 16 0.6834 0.1604 0.6834 0.8267
No log 0.3 18 0.6822 0.2121 0.6822 0.8260
No log 0.3333 20 0.6212 0.3499 0.6212 0.7882
No log 0.3667 22 0.6091 0.4388 0.6091 0.7804
No log 0.4 24 0.5567 0.4291 0.5567 0.7461
No log 0.4333 26 0.7828 0.4601 0.7828 0.8847
No log 0.4667 28 0.6911 0.5204 0.6911 0.8313
No log 0.5 30 0.5680 0.5212 0.5680 0.7537
No log 0.5333 32 1.0968 0.3171 1.0968 1.0473
No log 0.5667 34 1.1138 0.2923 1.1138 1.0554
No log 0.6 36 0.6948 0.4007 0.6948 0.8336
No log 0.6333 38 0.4674 0.6004 0.4674 0.6837
No log 0.6667 40 0.5827 0.5190 0.5827 0.7633
No log 0.7 42 0.4838 0.5601 0.4838 0.6956
No log 0.7333 44 0.4626 0.5265 0.4626 0.6802
No log 0.7667 46 0.6078 0.5101 0.6078 0.7796
No log 0.8 48 0.6182 0.5262 0.6182 0.7863
No log 0.8333 50 0.5112 0.4322 0.5112 0.7150
No log 0.8667 52 0.5217 0.5310 0.5217 0.7223
No log 0.9 54 0.7456 0.4844 0.7456 0.8635
No log 0.9333 56 0.6398 0.5468 0.6398 0.7999
No log 0.9667 58 0.4561 0.6109 0.4561 0.6754
No log 1.0 60 0.4673 0.5208 0.4673 0.6836
No log 1.0333 62 0.6208 0.5065 0.6208 0.7879
No log 1.0667 64 0.6061 0.5138 0.6061 0.7785
No log 1.1 66 0.5822 0.5312 0.5822 0.7630
No log 1.1333 68 0.6019 0.5599 0.6019 0.7758
No log 1.1667 70 0.4928 0.6620 0.4928 0.7020
No log 1.2 72 0.4555 0.6725 0.4555 0.6749
No log 1.2333 74 0.4588 0.6540 0.4588 0.6774
No log 1.2667 76 0.5074 0.6526 0.5074 0.7123
No log 1.3 78 0.7644 0.4208 0.7644 0.8743
No log 1.3333 80 0.6878 0.3978 0.6878 0.8293
No log 1.3667 82 0.5368 0.4949 0.5368 0.7327
No log 1.4 84 0.4631 0.6154 0.4631 0.6805
No log 1.4333 86 0.4701 0.6154 0.4701 0.6856
No log 1.4667 88 0.5077 0.5528 0.5077 0.7125
No log 1.5 90 0.4685 0.5985 0.4685 0.6844
No log 1.5333 92 0.4873 0.6341 0.4873 0.6980
No log 1.5667 94 0.6333 0.5408 0.6333 0.7958
No log 1.6 96 0.4806 0.5897 0.4806 0.6933
No log 1.6333 98 0.5361 0.5778 0.5361 0.7322
No log 1.6667 100 0.8300 0.4033 0.8300 0.9110
No log 1.7 102 0.7511 0.4413 0.7511 0.8666
No log 1.7333 104 0.4766 0.5528 0.4766 0.6904
No log 1.7667 106 0.4467 0.6186 0.4467 0.6684
No log 1.8 108 0.6802 0.5234 0.6802 0.8248
No log 1.8333 110 0.7963 0.4768 0.7963 0.8923
No log 1.8667 112 0.6273 0.4438 0.6273 0.7920
No log 1.9 114 0.4588 0.6045 0.4588 0.6774
No log 1.9333 116 0.4984 0.5811 0.4984 0.7060
No log 1.9667 118 0.5469 0.5614 0.5469 0.7395
No log 2.0 120 0.4526 0.6295 0.4526 0.6728
No log 2.0333 122 0.4422 0.6541 0.4422 0.6650
No log 2.0667 124 0.4387 0.7033 0.4387 0.6623
No log 2.1 126 0.4339 0.6648 0.4339 0.6587
No log 2.1333 128 0.4353 0.6556 0.4353 0.6598
No log 2.1667 130 0.4286 0.6648 0.4286 0.6547
No log 2.2 132 0.4362 0.6313 0.4362 0.6605
No log 2.2333 134 0.4454 0.6775 0.4454 0.6674
No log 2.2667 136 0.4165 0.6908 0.4165 0.6453
No log 2.3 138 0.4603 0.5956 0.4603 0.6784
No log 2.3333 140 0.5451 0.6065 0.5451 0.7383
No log 2.3667 142 0.5100 0.6148 0.5100 0.7141
No log 2.4 144 0.4376 0.7131 0.4376 0.6615
No log 2.4333 146 0.4219 0.7313 0.4219 0.6495
No log 2.4667 148 0.4102 0.6894 0.4102 0.6405
No log 2.5 150 0.4105 0.6542 0.4105 0.6407
No log 2.5333 152 0.4091 0.6747 0.4091 0.6396
No log 2.5667 154 0.4134 0.6747 0.4134 0.6430
No log 2.6 156 0.4504 0.6599 0.4504 0.6711
No log 2.6333 158 0.4800 0.6189 0.4800 0.6928
No log 2.6667 160 0.4391 0.6651 0.4391 0.6626
No log 2.7 162 0.5701 0.5599 0.5701 0.7550
No log 2.7333 164 0.5935 0.5249 0.5935 0.7704
No log 2.7667 166 0.5125 0.5869 0.5125 0.7159
No log 2.8 168 0.4648 0.6307 0.4648 0.6818
No log 2.8333 170 0.4625 0.6307 0.4625 0.6801
No log 2.8667 172 0.4587 0.6252 0.4587 0.6772
No log 2.9 174 0.4863 0.5655 0.4863 0.6974
No log 2.9333 176 0.4575 0.4984 0.4575 0.6764
No log 2.9667 178 0.4536 0.5815 0.4536 0.6735
No log 3.0 180 0.4631 0.6852 0.4631 0.6805
No log 3.0333 182 0.5064 0.5945 0.5064 0.7116
No log 3.0667 184 0.7636 0.5139 0.7636 0.8738
No log 3.1 186 0.8837 0.5123 0.8837 0.9400
No log 3.1333 188 0.7193 0.5058 0.7193 0.8481
No log 3.1667 190 0.5766 0.4504 0.5766 0.7593
No log 3.2 192 0.4443 0.6517 0.4443 0.6665
No log 3.2333 194 0.5240 0.6154 0.5240 0.7239
No log 3.2667 196 0.5580 0.4842 0.5580 0.7470
No log 3.3 198 0.4453 0.6305 0.4453 0.6673
No log 3.3333 200 0.4447 0.6082 0.4447 0.6669
No log 3.3667 202 0.4363 0.5897 0.4363 0.6605
No log 3.4 204 0.4324 0.6214 0.4324 0.6576
No log 3.4333 206 0.4399 0.6303 0.4399 0.6632
No log 3.4667 208 0.4724 0.5920 0.4724 0.6873
No log 3.5 210 0.4613 0.5741 0.4613 0.6792
No log 3.5333 212 0.4743 0.6214 0.4743 0.6887
No log 3.5667 214 0.4568 0.6118 0.4568 0.6759
No log 3.6 216 0.4786 0.6214 0.4786 0.6918
No log 3.6333 218 0.4540 0.6300 0.4540 0.6738
No log 3.6667 220 0.4195 0.6229 0.4195 0.6477
No log 3.7 222 0.4258 0.6068 0.4258 0.6526
No log 3.7333 224 0.4173 0.6053 0.4173 0.6460
No log 3.7667 226 0.4183 0.6422 0.4183 0.6467
No log 3.8 228 0.4213 0.6611 0.4213 0.6491
No log 3.8333 230 0.4072 0.6330 0.4072 0.6381
No log 3.8667 232 0.4212 0.6503 0.4212 0.6490
No log 3.9 234 0.4233 0.6503 0.4233 0.6506
No log 3.9333 236 0.4086 0.6330 0.4086 0.6392
No log 3.9667 238 0.4095 0.6542 0.4095 0.6399
No log 4.0 240 0.4189 0.6371 0.4189 0.6472
No log 4.0333 242 0.4371 0.6530 0.4371 0.6611
No log 4.0667 244 0.4277 0.6046 0.4277 0.6540
No log 4.1 246 0.4291 0.6125 0.4291 0.6551
No log 4.1333 248 0.4621 0.4816 0.4621 0.6798
No log 4.1667 250 0.4970 0.5498 0.4970 0.7050
No log 4.2 252 0.5522 0.5258 0.5522 0.7431
No log 4.2333 254 0.5284 0.4979 0.5284 0.7269
No log 4.2667 256 0.5183 0.5586 0.5183 0.7200
No log 4.3 258 0.5682 0.5773 0.5682 0.7538
No log 4.3333 260 0.5346 0.5362 0.5346 0.7312
No log 4.3667 262 0.5624 0.5570 0.5624 0.7499
No log 4.4 264 0.6704 0.4991 0.6704 0.8188
No log 4.4333 266 0.6074 0.5009 0.6074 0.7794
No log 4.4667 268 0.4935 0.5158 0.4935 0.7025
No log 4.5 270 0.4525 0.5209 0.4525 0.6727
No log 4.5333 272 0.4982 0.5158 0.4982 0.7059
No log 4.5667 274 0.6188 0.4853 0.6188 0.7866
No log 4.6 276 0.5399 0.4808 0.5399 0.7348
No log 4.6333 278 0.4211 0.5819 0.4211 0.6489
No log 4.6667 280 0.4099 0.6636 0.4099 0.6403
No log 4.7 282 0.4021 0.6422 0.4021 0.6341
No log 4.7333 284 0.4808 0.5481 0.4808 0.6934
No log 4.7667 286 0.4843 0.5435 0.4843 0.6959
No log 4.8 288 0.4230 0.6264 0.4230 0.6504
No log 4.8333 290 0.3918 0.6636 0.3918 0.6259
No log 4.8667 292 0.3907 0.6636 0.3907 0.6250
No log 4.9 294 0.3897 0.7133 0.3897 0.6242
No log 4.9333 296 0.4147 0.6604 0.4147 0.6439
No log 4.9667 298 0.4123 0.7140 0.4123 0.6421
No log 5.0 300 0.4416 0.7255 0.4416 0.6646
No log 5.0333 302 0.4916 0.6778 0.4916 0.7011
No log 5.0667 304 0.4670 0.6864 0.4670 0.6834
No log 5.1 306 0.4812 0.6367 0.4812 0.6937
No log 5.1333 308 0.5150 0.6310 0.5150 0.7176
No log 5.1667 310 0.5022 0.5455 0.5022 0.7087
No log 5.2 312 0.4347 0.5868 0.4347 0.6593
No log 5.2333 314 0.4117 0.6555 0.4117 0.6417
No log 5.2667 316 0.4193 0.6317 0.4193 0.6475
No log 5.3 318 0.4474 0.5715 0.4474 0.6689
No log 5.3333 320 0.4244 0.6096 0.4244 0.6514
No log 5.3667 322 0.4204 0.6326 0.4204 0.6484
No log 5.4 324 0.4074 0.6426 0.4074 0.6383
No log 5.4333 326 0.4068 0.6914 0.4068 0.6378
No log 5.4667 328 0.4170 0.6377 0.4170 0.6458
No log 5.5 330 0.4447 0.5999 0.4447 0.6669
No log 5.5333 332 0.4153 0.6542 0.4153 0.6445
No log 5.5667 334 0.4387 0.6289 0.4387 0.6623
No log 5.6 336 0.4454 0.5768 0.4454 0.6674
No log 5.6333 338 0.4340 0.5995 0.4340 0.6588
No log 5.6667 340 0.4520 0.6301 0.4520 0.6723
No log 5.7 342 0.4766 0.6688 0.4766 0.6904
No log 5.7333 344 0.4538 0.6564 0.4538 0.6737
No log 5.7667 346 0.5142 0.5553 0.5142 0.7171
No log 5.8 348 0.5597 0.5570 0.5597 0.7482
No log 5.8333 350 0.5605 0.5570 0.5605 0.7487
No log 5.8667 352 0.5614 0.5570 0.5614 0.7492
No log 5.9 354 0.5550 0.5123 0.5550 0.7450
No log 5.9333 356 0.5907 0.4870 0.5907 0.7686
No log 5.9667 358 0.6194 0.5103 0.6194 0.7870
No log 6.0 360 0.6256 0.4735 0.6256 0.7910
No log 6.0333 362 0.5374 0.5538 0.5374 0.7331
No log 6.0667 364 0.4674 0.6129 0.4674 0.6837
No log 6.1 366 0.4580 0.6643 0.4580 0.6767
No log 6.1333 368 0.4731 0.5324 0.4731 0.6878
No log 6.1667 370 0.5379 0.5659 0.5379 0.7334
No log 6.2 372 0.5143 0.5140 0.5143 0.7172
No log 6.2333 374 0.4674 0.5708 0.4674 0.6837
No log 6.2667 376 0.4620 0.5719 0.4620 0.6797
No log 6.3 378 0.4743 0.5166 0.4743 0.6887
No log 6.3333 380 0.4869 0.5223 0.4869 0.6978
No log 6.3667 382 0.5087 0.5659 0.5087 0.7132
No log 6.4 384 0.4915 0.5223 0.4915 0.7010
No log 6.4333 386 0.4554 0.6214 0.4554 0.6748
No log 6.4667 388 0.4582 0.5945 0.4582 0.6769
No log 6.5 390 0.4836 0.5733 0.4836 0.6954
No log 6.5333 392 0.4642 0.6020 0.4642 0.6813
No log 6.5667 394 0.4520 0.6564 0.4520 0.6723
No log 6.6 396 0.4503 0.5831 0.4503 0.6710
No log 6.6333 398 0.4651 0.5151 0.4651 0.6820
No log 6.6667 400 0.4941 0.4795 0.4941 0.7029
No log 6.7 402 0.5129 0.5271 0.5129 0.7162
No log 6.7333 404 0.4811 0.4888 0.4811 0.6936
No log 6.7667 406 0.4790 0.5341 0.4790 0.6921
No log 6.8 408 0.4792 0.5711 0.4792 0.6922
No log 6.8333 410 0.4822 0.5742 0.4822 0.6944
No log 6.8667 412 0.4818 0.5272 0.4818 0.6941
No log 6.9 414 0.4845 0.5123 0.4845 0.6961
No log 6.9333 416 0.4914 0.4888 0.4914 0.7010
No log 6.9667 418 0.5101 0.4243 0.5101 0.7142
No log 7.0 420 0.5494 0.4243 0.5494 0.7412
No log 7.0333 422 0.5714 0.5091 0.5714 0.7559
No log 7.0667 424 0.5478 0.4597 0.5478 0.7402
No log 7.1 426 0.5104 0.4291 0.5104 0.7144
No log 7.1333 428 0.4988 0.4569 0.4988 0.7063
No log 7.1667 430 0.4971 0.4569 0.4971 0.7050
No log 7.2 432 0.4981 0.4291 0.4981 0.7057
No log 7.2333 434 0.5370 0.5158 0.5370 0.7328
No log 7.2667 436 0.5539 0.5067 0.5539 0.7442
No log 7.3 438 0.5021 0.5252 0.5021 0.7086
No log 7.3333 440 0.4755 0.4847 0.4755 0.6896
No log 7.3667 442 0.4698 0.4955 0.4698 0.6855
No log 7.4 444 0.4729 0.5840 0.4729 0.6877
No log 7.4333 446 0.4842 0.5767 0.4842 0.6958
No log 7.4667 448 0.4704 0.5840 0.4704 0.6859
No log 7.5 450 0.4967 0.5485 0.4967 0.7048
No log 7.5333 452 0.5344 0.4997 0.5344 0.7311
No log 7.5667 454 0.5278 0.4835 0.5278 0.7265
No log 7.6 456 0.5535 0.4753 0.5535 0.7439
No log 7.6333 458 0.5487 0.5189 0.5487 0.7407
No log 7.6667 460 0.5688 0.5189 0.5688 0.7542
No log 7.7 462 0.5742 0.5189 0.5742 0.7578
No log 7.7333 464 0.5610 0.5189 0.5610 0.7490
No log 7.7667 466 0.5205 0.5017 0.5205 0.7215
No log 7.8 468 0.4852 0.5815 0.4852 0.6965
No log 7.8333 470 0.5153 0.4653 0.5153 0.7178
No log 7.8667 472 0.5499 0.5119 0.5499 0.7416
No log 7.9 474 0.5610 0.4833 0.5610 0.7490
No log 7.9333 476 0.5037 0.5065 0.5037 0.7097
No log 7.9667 478 0.4548 0.5600 0.4548 0.6744
No log 8.0 480 0.4998 0.5639 0.4998 0.7070
No log 8.0333 482 0.5397 0.4905 0.5397 0.7347
No log 8.0667 484 0.5300 0.5330 0.5300 0.7280
No log 8.1 486 0.5073 0.5448 0.5073 0.7123
No log 8.1333 488 0.4890 0.5056 0.4890 0.6993
No log 8.1667 490 0.4887 0.5614 0.4887 0.6991
No log 8.2 492 0.4897 0.5655 0.4897 0.6998
No log 8.2333 494 0.5021 0.5553 0.5021 0.7086
No log 8.2667 496 0.4799 0.6187 0.4799 0.6927
No log 8.3 498 0.4477 0.6198 0.4477 0.6691
0.2831 8.3333 500 0.4513 0.5826 0.4513 0.6718
0.2831 8.3667 502 0.4532 0.5687 0.4532 0.6732
0.2831 8.4 504 0.4574 0.5507 0.4574 0.6763
0.2831 8.4333 506 0.4688 0.6228 0.4688 0.6847
0.2831 8.4667 508 0.4916 0.5485 0.4916 0.7012
0.2831 8.5 510 0.5003 0.5254 0.5003 0.7073
0.2831 8.5333 512 0.4863 0.5752 0.4863 0.6974
0.2831 8.5667 514 0.4643 0.5396 0.4643 0.6814
0.2831 8.6 516 0.4657 0.5003 0.4657 0.6824

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k12_task7_organization

Finetuned
(4019)
this model