ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k20_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4454
  • Qwk: 0.6426
  • Mse: 0.4454
  • Rmse: 0.6674

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.02 2 2.7161 -0.0593 2.7161 1.6481
No log 0.04 4 1.3498 0.1947 1.3498 1.1618
No log 0.06 6 0.6855 0.1372 0.6855 0.8279
No log 0.08 8 0.7684 0.3630 0.7684 0.8766
No log 0.1 10 1.0217 0.2903 1.0217 1.0108
No log 0.12 12 0.6671 0.4251 0.6671 0.8167
No log 0.14 14 0.6452 0.5016 0.6452 0.8032
No log 0.16 16 0.6868 0.4833 0.6868 0.8287
No log 0.18 18 0.6235 0.4879 0.6235 0.7896
No log 0.2 20 0.5132 0.5152 0.5132 0.7164
No log 0.22 22 0.5156 0.5361 0.5156 0.7181
No log 0.24 24 0.5549 0.4997 0.5549 0.7449
No log 0.26 26 0.5915 0.5143 0.5915 0.7691
No log 0.28 28 0.5787 0.5057 0.5787 0.7608
No log 0.3 30 0.5284 0.6943 0.5284 0.7269
No log 0.32 32 0.5133 0.6855 0.5133 0.7165
No log 0.34 34 0.7070 0.4874 0.7070 0.8408
No log 0.36 36 0.7536 0.4699 0.7536 0.8681
No log 0.38 38 0.6138 0.5254 0.6138 0.7835
No log 0.4 40 0.5064 0.6410 0.5064 0.7116
No log 0.42 42 0.4596 0.6657 0.4596 0.6779
No log 0.44 44 0.4491 0.6307 0.4491 0.6701
No log 0.46 46 0.4299 0.6847 0.4299 0.6557
No log 0.48 48 0.4782 0.5945 0.4782 0.6915
No log 0.5 50 0.5257 0.5900 0.5257 0.7251
No log 0.52 52 0.4530 0.6457 0.4530 0.6731
No log 0.54 54 0.4289 0.6735 0.4289 0.6549
No log 0.56 56 0.4535 0.6922 0.4535 0.6734
No log 0.58 58 0.5425 0.6170 0.5425 0.7366
No log 0.6 60 0.5732 0.6494 0.5732 0.7571
No log 0.62 62 0.5071 0.7267 0.5071 0.7121
No log 0.64 64 0.5620 0.6756 0.5620 0.7497
No log 0.66 66 0.5252 0.6698 0.5252 0.7247
No log 0.68 68 0.5144 0.7408 0.5144 0.7172
No log 0.7 70 0.5709 0.6641 0.5709 0.7556
No log 0.72 72 0.5406 0.6645 0.5406 0.7353
No log 0.74 74 0.4358 0.6796 0.4358 0.6602
No log 0.76 76 0.4366 0.6455 0.4366 0.6608
No log 0.78 78 0.4531 0.6848 0.4531 0.6732
No log 0.8 80 0.4288 0.6382 0.4288 0.6549
No log 0.82 82 0.4276 0.7349 0.4276 0.6539
No log 0.84 84 0.4687 0.6930 0.4687 0.6847
No log 0.86 86 0.4600 0.7111 0.4600 0.6783
No log 0.88 88 0.4952 0.6868 0.4952 0.7037
No log 0.9 90 0.4717 0.7407 0.4717 0.6868
No log 0.92 92 0.4684 0.6788 0.4684 0.6844
No log 0.94 94 0.5613 0.6095 0.5613 0.7492
No log 0.96 96 0.4894 0.6632 0.4894 0.6996
No log 0.98 98 0.4906 0.6480 0.4906 0.7004
No log 1.0 100 0.4783 0.6477 0.4783 0.6916
No log 1.02 102 0.4937 0.5840 0.4937 0.7027
No log 1.04 104 0.4399 0.6469 0.4399 0.6632
No log 1.06 106 0.4438 0.6694 0.4438 0.6662
No log 1.08 108 0.4290 0.5812 0.4290 0.6550
No log 1.1 110 0.5196 0.5123 0.5196 0.7208
No log 1.12 112 0.6614 0.4921 0.6614 0.8132
No log 1.1400 114 0.5349 0.5354 0.5349 0.7314
No log 1.16 116 0.4163 0.6589 0.4163 0.6452
No log 1.18 118 0.5209 0.6055 0.5209 0.7217
No log 1.2 120 0.5326 0.5982 0.5326 0.7298
No log 1.22 122 0.4510 0.5939 0.4510 0.6716
No log 1.24 124 0.6275 0.4949 0.6275 0.7922
No log 1.26 126 0.7897 0.5132 0.7897 0.8887
No log 1.28 128 0.6212 0.4684 0.6212 0.7882
No log 1.3 130 0.5068 0.5327 0.5068 0.7119
No log 1.32 132 0.5284 0.5438 0.5284 0.7269
No log 1.34 134 0.5745 0.5935 0.5745 0.7580
No log 1.3600 136 0.4737 0.6210 0.4737 0.6882
No log 1.38 138 0.5064 0.5512 0.5064 0.7116
No log 1.4 140 0.5295 0.5512 0.5295 0.7277
No log 1.42 142 0.4909 0.6186 0.4909 0.7006
No log 1.44 144 0.4925 0.6423 0.4925 0.7018
No log 1.46 146 0.4835 0.6735 0.4835 0.6953
No log 1.48 148 0.4748 0.6059 0.4748 0.6891
No log 1.5 150 0.4603 0.5891 0.4603 0.6784
No log 1.52 152 0.4430 0.5753 0.4430 0.6656
No log 1.54 154 0.5297 0.4491 0.5297 0.7278
No log 1.56 156 0.6633 0.4907 0.6633 0.8144
No log 1.58 158 0.7904 0.4279 0.7904 0.8891
No log 1.6 160 0.6147 0.4925 0.6147 0.7841
No log 1.62 162 0.4298 0.6374 0.4298 0.6556
No log 1.6400 164 0.5162 0.5735 0.5162 0.7184
No log 1.6600 166 0.6983 0.4321 0.6983 0.8356
No log 1.6800 168 0.6321 0.4799 0.6321 0.7950
No log 1.7 170 0.4548 0.6793 0.4548 0.6744
No log 1.72 172 0.4868 0.6070 0.4868 0.6977
No log 1.74 174 0.5603 0.5895 0.5603 0.7486
No log 1.76 176 0.5921 0.5986 0.5921 0.7695
No log 1.78 178 0.4809 0.5557 0.4809 0.6935
No log 1.8 180 0.4906 0.6158 0.4906 0.7004
No log 1.8200 182 0.5565 0.6390 0.5565 0.7460
No log 1.8400 184 0.5156 0.6379 0.5156 0.7180
No log 1.8600 186 0.5091 0.5391 0.5091 0.7135
No log 1.88 188 0.5314 0.5614 0.5314 0.7290
No log 1.9 190 0.6055 0.5595 0.6055 0.7781
No log 1.92 192 0.5914 0.5561 0.5914 0.7690
No log 1.94 194 0.5603 0.5276 0.5603 0.7486
No log 1.96 196 0.5101 0.4772 0.5101 0.7142
No log 1.98 198 0.5038 0.4986 0.5038 0.7098
No log 2.0 200 0.5162 0.5447 0.5162 0.7185
No log 2.02 202 0.5244 0.5447 0.5244 0.7241
No log 2.04 204 0.5057 0.5770 0.5057 0.7111
No log 2.06 206 0.5023 0.4678 0.5023 0.7087
No log 2.08 208 0.5095 0.5505 0.5095 0.7138
No log 2.1 210 0.5285 0.5786 0.5285 0.7270
No log 2.12 212 0.5482 0.5942 0.5482 0.7404
No log 2.14 214 0.5276 0.5696 0.5276 0.7264
No log 2.16 216 0.5484 0.5748 0.5484 0.7405
No log 2.18 218 0.4858 0.5729 0.4858 0.6970
No log 2.2 220 0.4431 0.6371 0.4431 0.6657
No log 2.22 222 0.4439 0.6289 0.4439 0.6663
No log 2.24 224 0.4715 0.6318 0.4715 0.6866
No log 2.26 226 0.5149 0.6167 0.5149 0.7176
No log 2.2800 228 0.5800 0.5810 0.5800 0.7616
No log 2.3 230 0.5830 0.6054 0.5830 0.7635
No log 2.32 232 0.5528 0.5927 0.5528 0.7435
No log 2.34 234 0.5027 0.5761 0.5027 0.7090
No log 2.36 236 0.4977 0.6188 0.4977 0.7054
No log 2.38 238 0.4390 0.6643 0.4390 0.6626
No log 2.4 240 0.4358 0.6073 0.4358 0.6601
No log 2.42 242 0.4517 0.6158 0.4517 0.6721
No log 2.44 244 0.4703 0.6100 0.4703 0.6858
No log 2.46 246 0.4525 0.6405 0.4525 0.6727
No log 2.48 248 0.5417 0.6322 0.5417 0.7360
No log 2.5 250 0.5881 0.6128 0.5881 0.7669
No log 2.52 252 0.5600 0.5962 0.5600 0.7483
No log 2.54 254 0.5424 0.6311 0.5424 0.7365
No log 2.56 256 0.4910 0.6143 0.4910 0.7007
No log 2.58 258 0.4761 0.5872 0.4761 0.6900
No log 2.6 260 0.5192 0.5237 0.5192 0.7205
No log 2.62 262 0.5055 0.5111 0.5055 0.7110
No log 2.64 264 0.4772 0.5065 0.4772 0.6908
No log 2.66 266 0.4724 0.4869 0.4724 0.6873
No log 2.68 268 0.4483 0.5538 0.4483 0.6695
No log 2.7 270 0.4619 0.5617 0.4619 0.6796
No log 2.7200 272 0.4915 0.6152 0.4915 0.7011
No log 2.74 274 0.4974 0.6165 0.4974 0.7052
No log 2.76 276 0.5039 0.5712 0.5039 0.7099
No log 2.7800 278 0.4816 0.5786 0.4816 0.6940
No log 2.8 280 0.5329 0.5845 0.5329 0.7300
No log 2.82 282 0.4744 0.5786 0.4744 0.6887
No log 2.84 284 0.4362 0.5784 0.4362 0.6605
No log 2.86 286 0.4831 0.5687 0.4831 0.6950
No log 2.88 288 0.4839 0.5619 0.4839 0.6956
No log 2.9 290 0.4280 0.6371 0.4280 0.6542
No log 2.92 292 0.4333 0.6566 0.4333 0.6583
No log 2.94 294 0.4384 0.6566 0.4384 0.6621
No log 2.96 296 0.4599 0.6032 0.4599 0.6781
No log 2.98 298 0.5209 0.5524 0.5209 0.7217
No log 3.0 300 0.5715 0.5587 0.5715 0.7560
No log 3.02 302 0.5875 0.5620 0.5875 0.7665
No log 3.04 304 0.5864 0.5546 0.5864 0.7658
No log 3.06 306 0.5177 0.5559 0.5177 0.7195
No log 3.08 308 0.4926 0.5566 0.4926 0.7018
No log 3.1 310 0.4784 0.5612 0.4784 0.6916
No log 3.12 312 0.5119 0.5353 0.5119 0.7154
No log 3.14 314 0.5191 0.5418 0.5191 0.7205
No log 3.16 316 0.5452 0.5524 0.5452 0.7384
No log 3.18 318 0.4862 0.5827 0.4862 0.6973
No log 3.2 320 0.4549 0.5707 0.4549 0.6745
No log 3.22 322 0.4455 0.6371 0.4455 0.6675
No log 3.24 324 0.4434 0.6554 0.4434 0.6659
No log 3.26 326 0.4619 0.5874 0.4619 0.6796
No log 3.2800 328 0.5127 0.5487 0.5127 0.7160
No log 3.3 330 0.5162 0.5646 0.5162 0.7185
No log 3.32 332 0.5031 0.5827 0.5031 0.7093
No log 3.34 334 0.5041 0.5669 0.5041 0.7100
No log 3.36 336 0.5072 0.5174 0.5072 0.7122
No log 3.38 338 0.5343 0.5729 0.5343 0.7310
No log 3.4 340 0.6013 0.5455 0.6013 0.7754
No log 3.42 342 0.6603 0.52 0.6603 0.8126
No log 3.44 344 0.7127 0.5124 0.7127 0.8442
No log 3.46 346 0.6199 0.5339 0.6199 0.7873
No log 3.48 348 0.4777 0.5784 0.4777 0.6912
No log 3.5 350 0.4464 0.5913 0.4464 0.6682
No log 3.52 352 0.4498 0.5711 0.4498 0.6707
No log 3.54 354 0.4778 0.5742 0.4778 0.6912
No log 3.56 356 0.4720 0.5515 0.4720 0.6870
No log 3.58 358 0.4329 0.5913 0.4329 0.6580
No log 3.6 360 0.4228 0.6277 0.4228 0.6503
No log 3.62 362 0.4192 0.6939 0.4192 0.6474
No log 3.64 364 0.4221 0.6939 0.4221 0.6497
No log 3.66 366 0.4607 0.6061 0.4607 0.6787
No log 3.68 368 0.5429 0.5543 0.5429 0.7368
No log 3.7 370 0.6029 0.5103 0.6029 0.7765
No log 3.7200 372 0.5422 0.5723 0.5422 0.7364
No log 3.74 374 0.4471 0.6783 0.4471 0.6687
No log 3.76 376 0.4463 0.6265 0.4463 0.6681
No log 3.7800 378 0.4598 0.6173 0.4598 0.6781
No log 3.8 380 0.4797 0.5447 0.4797 0.6926
No log 3.82 382 0.4646 0.5580 0.4646 0.6816
No log 3.84 384 0.4345 0.6171 0.4345 0.6591
No log 3.86 386 0.4289 0.7381 0.4289 0.6549
No log 3.88 388 0.4922 0.6637 0.4922 0.7016
No log 3.9 390 0.5268 0.5779 0.5268 0.7258
No log 3.92 392 0.4862 0.7059 0.4862 0.6973
No log 3.94 394 0.4446 0.7430 0.4446 0.6668
No log 3.96 396 0.4341 0.6891 0.4341 0.6588
No log 3.98 398 0.4379 0.6484 0.4379 0.6618
No log 4.0 400 0.4481 0.6307 0.4481 0.6694
No log 4.02 402 0.4516 0.6009 0.4516 0.6720
No log 4.04 404 0.4669 0.6083 0.4669 0.6833
No log 4.06 406 0.4730 0.6346 0.4730 0.6877
No log 4.08 408 0.4793 0.5974 0.4793 0.6923
No log 4.1 410 0.5005 0.5741 0.5005 0.7074
No log 4.12 412 0.5203 0.5081 0.5203 0.7213
No log 4.14 414 0.5104 0.5308 0.5104 0.7144
No log 4.16 416 0.4920 0.5254 0.4920 0.7015
No log 4.18 418 0.5102 0.5308 0.5102 0.7143
No log 4.2 420 0.5843 0.5813 0.5843 0.7644
No log 4.22 422 0.5965 0.5695 0.5965 0.7723
No log 4.24 424 0.5031 0.6035 0.5031 0.7093
No log 4.26 426 0.4563 0.6973 0.4563 0.6755
No log 4.28 428 0.4889 0.6803 0.4889 0.6992
No log 4.3 430 0.4739 0.6803 0.4739 0.6884
No log 4.32 432 0.4405 0.7475 0.4405 0.6637
No log 4.34 434 0.4249 0.6541 0.4249 0.6519
No log 4.36 436 0.4536 0.6083 0.4536 0.6735
No log 4.38 438 0.4462 0.6083 0.4462 0.6680
No log 4.4 440 0.4422 0.6083 0.4422 0.6649
No log 4.42 442 0.4295 0.6346 0.4295 0.6554
No log 4.44 444 0.4283 0.6346 0.4283 0.6545
No log 4.46 446 0.4422 0.6346 0.4422 0.6650
No log 4.48 448 0.4507 0.6353 0.4507 0.6713
No log 4.5 450 0.4531 0.6353 0.4531 0.6731
No log 4.52 452 0.4684 0.6705 0.4684 0.6844
No log 4.54 454 0.5167 0.5677 0.5167 0.7188
No log 4.5600 456 0.5577 0.5624 0.5577 0.7468
No log 4.58 458 0.5494 0.5591 0.5494 0.7412
No log 4.6 460 0.4879 0.5852 0.4879 0.6985
No log 4.62 462 0.4357 0.6634 0.4357 0.6601
No log 4.64 464 0.4426 0.7022 0.4426 0.6653
No log 4.66 466 0.4442 0.7192 0.4442 0.6665
No log 4.68 468 0.4385 0.7069 0.4385 0.6622
No log 4.7 470 0.4295 0.6637 0.4295 0.6553
No log 4.72 472 0.4413 0.6777 0.4413 0.6643
No log 4.74 474 0.4348 0.6612 0.4348 0.6594
No log 4.76 476 0.4701 0.6201 0.4701 0.6856
No log 4.78 478 0.4739 0.5736 0.4739 0.6884
No log 4.8 480 0.4609 0.6156 0.4609 0.6789
No log 4.82 482 0.4306 0.6303 0.4306 0.6562
No log 4.84 484 0.4001 0.6736 0.4001 0.6326
No log 4.86 486 0.4023 0.7012 0.4023 0.6343
No log 4.88 488 0.4120 0.7273 0.4120 0.6418
No log 4.9 490 0.4233 0.6735 0.4233 0.6506
No log 4.92 492 0.4344 0.7910 0.4344 0.6591
No log 4.9400 494 0.4402 0.7668 0.4402 0.6635
No log 4.96 496 0.4648 0.7031 0.4648 0.6818
No log 4.98 498 0.4918 0.6965 0.4918 0.7013
0.2959 5.0 500 0.4854 0.7124 0.4854 0.6967
0.2959 5.02 502 0.4442 0.7110 0.4442 0.6665
0.2959 5.04 504 0.4444 0.6254 0.4444 0.6666
0.2959 5.06 506 0.4627 0.6228 0.4627 0.6802
0.2959 5.08 508 0.4537 0.6426 0.4537 0.6736
0.2959 5.1 510 0.4454 0.6426 0.4454 0.6674

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k20_task7_organization

Finetuned
(4019)
this model