ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k20_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5460
  • Qwk: 0.4774
  • Mse: 0.5460
  • Rmse: 0.7389

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.02 2 2.4336 -0.0262 2.4336 1.5600
No log 0.04 4 1.1177 0.1251 1.1177 1.0572
No log 0.06 6 0.7027 0.1321 0.7027 0.8383
No log 0.08 8 0.6669 0.1846 0.6669 0.8166
No log 0.1 10 0.8837 0.3455 0.8837 0.9400
No log 0.12 12 1.2123 0.1283 1.2123 1.1011
No log 0.14 14 1.0419 0.1621 1.0419 1.0207
No log 0.16 16 0.9212 0.2183 0.9212 0.9598
No log 0.18 18 0.9159 0.2316 0.9159 0.9570
No log 0.2 20 0.8530 0.3819 0.8530 0.9236
No log 0.22 22 0.7369 0.1617 0.7369 0.8584
No log 0.24 24 0.6641 0.0937 0.6641 0.8149
No log 0.26 26 0.6216 0.2142 0.6216 0.7884
No log 0.28 28 0.5442 0.4314 0.5442 0.7377
No log 0.3 30 0.5884 0.3831 0.5884 0.7671
No log 0.32 32 0.6476 0.5083 0.6476 0.8047
No log 0.34 34 0.5482 0.4888 0.5482 0.7404
No log 0.36 36 0.4813 0.6052 0.4813 0.6937
No log 0.38 38 0.5166 0.5498 0.5166 0.7187
No log 0.4 40 0.5304 0.5877 0.5304 0.7283
No log 0.42 42 0.8805 0.3262 0.8805 0.9383
No log 0.44 44 0.8463 0.3322 0.8463 0.9200
No log 0.46 46 0.5334 0.5452 0.5334 0.7304
No log 0.48 48 0.6184 0.5481 0.6184 0.7864
No log 0.5 50 0.8713 0.3790 0.8713 0.9334
No log 0.52 52 0.6121 0.5378 0.6121 0.7824
No log 0.54 54 0.4697 0.5617 0.4697 0.6854
No log 0.56 56 0.6125 0.5152 0.6125 0.7826
No log 0.58 58 0.6373 0.4589 0.6373 0.7983
No log 0.6 60 0.5841 0.4322 0.5841 0.7642
No log 0.62 62 0.5709 0.4662 0.5709 0.7556
No log 0.64 64 0.8000 0.4951 0.8000 0.8944
No log 0.66 66 1.0245 0.4011 1.0245 1.0122
No log 0.68 68 0.7877 0.5310 0.7877 0.8875
No log 0.7 70 0.5326 0.4596 0.5326 0.7298
No log 0.72 72 0.5103 0.4448 0.5103 0.7144
No log 0.74 74 0.5131 0.4561 0.5131 0.7163
No log 0.76 76 0.4900 0.4838 0.4900 0.7000
No log 0.78 78 0.4794 0.5457 0.4794 0.6924
No log 0.8 80 0.5166 0.5571 0.5166 0.7187
No log 0.82 82 0.5016 0.5798 0.5016 0.7083
No log 0.84 84 0.5150 0.6353 0.5150 0.7176
No log 0.86 86 0.6014 0.4638 0.6014 0.7755
No log 0.88 88 0.6618 0.4794 0.6618 0.8135
No log 0.9 90 0.5386 0.6222 0.5386 0.7339
No log 0.92 92 0.6051 0.5765 0.6051 0.7779
No log 0.94 94 0.6192 0.6367 0.6192 0.7869
No log 0.96 96 0.6490 0.5742 0.6490 0.8056
No log 0.98 98 0.8902 0.4576 0.8902 0.9435
No log 1.0 100 0.9072 0.4386 0.9072 0.9525
No log 1.02 102 0.6718 0.5213 0.6718 0.8196
No log 1.04 104 0.5507 0.5589 0.5507 0.7421
No log 1.06 106 0.5274 0.5476 0.5274 0.7262
No log 1.08 108 0.5014 0.5231 0.5014 0.7081
No log 1.1 110 0.4889 0.6275 0.4889 0.6992
No log 1.12 112 0.4941 0.5226 0.4941 0.7029
No log 1.1400 114 0.5404 0.5576 0.5404 0.7351
No log 1.16 116 0.5000 0.5827 0.5000 0.7071
No log 1.18 118 0.5057 0.5705 0.5057 0.7111
No log 1.2 120 0.4649 0.6554 0.4649 0.6818
No log 1.22 122 0.5905 0.5415 0.5905 0.7685
No log 1.24 124 0.6216 0.5474 0.6216 0.7884
No log 1.26 126 0.5379 0.5452 0.5379 0.7334
No log 1.28 128 0.5202 0.5150 0.5202 0.7213
No log 1.3 130 0.5333 0.5034 0.5333 0.7303
No log 1.32 132 0.5977 0.5455 0.5977 0.7731
No log 1.34 134 0.6683 0.5508 0.6683 0.8175
No log 1.3600 136 0.5604 0.5291 0.5604 0.7486
No log 1.38 138 0.5166 0.4473 0.5166 0.7188
No log 1.4 140 0.4885 0.4857 0.4885 0.6989
No log 1.42 142 0.5161 0.4507 0.5161 0.7184
No log 1.44 144 0.5903 0.5687 0.5903 0.7683
No log 1.46 146 0.5110 0.4997 0.5110 0.7149
No log 1.48 148 0.5085 0.5722 0.5085 0.7131
No log 1.5 150 0.6388 0.5970 0.6388 0.7992
No log 1.52 152 0.5558 0.5927 0.5558 0.7455
No log 1.54 154 0.4846 0.4954 0.4846 0.6961
No log 1.56 156 0.5322 0.5599 0.5322 0.7295
No log 1.58 158 0.7361 0.4338 0.7361 0.8579
No log 1.6 160 0.7054 0.4815 0.7054 0.8399
No log 1.62 162 0.5173 0.6181 0.5173 0.7192
No log 1.6400 164 0.5233 0.6189 0.5233 0.7234
No log 1.6600 166 0.8778 0.5501 0.8778 0.9369
No log 1.6800 168 0.9912 0.4633 0.9912 0.9956
No log 1.7 170 0.8150 0.5157 0.8150 0.9028
No log 1.72 172 0.5731 0.5857 0.5731 0.7570
No log 1.74 174 0.4673 0.5413 0.4673 0.6836
No log 1.76 176 0.5112 0.5868 0.5112 0.7150
No log 1.78 178 0.5368 0.5678 0.5368 0.7327
No log 1.8 180 0.5065 0.6101 0.5065 0.7117
No log 1.8200 182 0.5677 0.5161 0.5677 0.7535
No log 1.8400 184 0.6485 0.5251 0.6485 0.8053
No log 1.8600 186 0.6607 0.5062 0.6607 0.8128
No log 1.88 188 0.6159 0.5323 0.6159 0.7848
No log 1.9 190 0.5391 0.4872 0.5391 0.7342
No log 1.92 192 0.5168 0.5133 0.5168 0.7189
No log 1.94 194 0.5070 0.5061 0.5070 0.7120
No log 1.96 196 0.5092 0.5042 0.5092 0.7136
No log 1.98 198 0.5705 0.4582 0.5705 0.7553
No log 2.0 200 0.6946 0.5310 0.6946 0.8334
No log 2.02 202 0.6341 0.5124 0.6341 0.7963
No log 2.04 204 0.5996 0.6131 0.5996 0.7743
No log 2.06 206 0.6785 0.5581 0.6785 0.8237
No log 2.08 208 0.5603 0.5863 0.5603 0.7485
No log 2.1 210 0.4828 0.6242 0.4828 0.6948
No log 2.12 212 0.4738 0.5397 0.4738 0.6883
No log 2.14 214 0.4622 0.5003 0.4622 0.6798
No log 2.16 216 0.4613 0.4895 0.4613 0.6792
No log 2.18 218 0.5422 0.5609 0.5422 0.7364
No log 2.2 220 0.5679 0.5609 0.5679 0.7536
No log 2.22 222 0.5358 0.5373 0.5358 0.7320
No log 2.24 224 0.4739 0.5715 0.4739 0.6884
No log 2.26 226 0.4894 0.5479 0.4894 0.6996
No log 2.2800 228 0.4965 0.4762 0.4965 0.7046
No log 2.3 230 0.5071 0.4888 0.5071 0.7121
No log 2.32 232 0.6275 0.5085 0.6275 0.7922
No log 2.34 234 0.6429 0.4933 0.6429 0.8018
No log 2.36 236 0.5460 0.5131 0.5460 0.7389
No log 2.38 238 0.4982 0.4962 0.4982 0.7059
No log 2.4 240 0.5455 0.3754 0.5455 0.7386
No log 2.42 242 0.5730 0.4204 0.5730 0.7569
No log 2.44 244 0.5306 0.5127 0.5306 0.7284
No log 2.46 246 0.4739 0.5286 0.4739 0.6884
No log 2.48 248 0.4972 0.5125 0.4972 0.7051
No log 2.5 250 0.5733 0.5692 0.5733 0.7572
No log 2.52 252 0.5502 0.5814 0.5502 0.7418
No log 2.54 254 0.5114 0.5190 0.5114 0.7151
No log 2.56 256 0.4824 0.4828 0.4824 0.6946
No log 2.58 258 0.4908 0.4945 0.4908 0.7006
No log 2.6 260 0.5071 0.5233 0.5071 0.7121
No log 2.62 262 0.5454 0.4959 0.5454 0.7385
No log 2.64 264 0.5076 0.4782 0.5076 0.7125
No log 2.66 266 0.4674 0.5703 0.4674 0.6837
No log 2.68 268 0.4661 0.5446 0.4661 0.6827
No log 2.7 270 0.4709 0.5943 0.4709 0.6862
No log 2.7200 272 0.5029 0.5822 0.5029 0.7091
No log 2.74 274 0.4940 0.6143 0.4940 0.7029
No log 2.76 276 0.4947 0.6052 0.4947 0.7034
No log 2.7800 278 0.5029 0.6052 0.5029 0.7092
No log 2.8 280 0.5189 0.6341 0.5189 0.7204
No log 2.82 282 0.5380 0.5884 0.5380 0.7335
No log 2.84 284 0.5818 0.5096 0.5818 0.7627
No log 2.86 286 0.5932 0.4893 0.5932 0.7702
No log 2.88 288 0.6177 0.4759 0.6177 0.7859
No log 2.9 290 0.6082 0.4964 0.6082 0.7798
No log 2.92 292 0.6086 0.4964 0.6086 0.7801
No log 2.94 294 0.5151 0.5086 0.5151 0.7177
No log 2.96 296 0.4882 0.5327 0.4882 0.6987
No log 2.98 298 0.4704 0.5034 0.4704 0.6858
No log 3.0 300 0.4879 0.5086 0.4879 0.6985
No log 3.02 302 0.5181 0.5712 0.5181 0.7198
No log 3.04 304 0.4900 0.5682 0.4900 0.7000
No log 3.06 306 0.4720 0.6382 0.4720 0.6870
No log 3.08 308 0.5000 0.5524 0.5000 0.7071
No log 3.1 310 0.5000 0.5592 0.5000 0.7071
No log 3.12 312 0.4885 0.5936 0.4885 0.6989
No log 3.14 314 0.5164 0.5326 0.5164 0.7186
No log 3.16 316 0.6109 0.5190 0.6109 0.7816
No log 3.18 318 0.6291 0.4795 0.6291 0.7932
No log 3.2 320 0.6128 0.4884 0.6128 0.7828
No log 3.22 322 0.6407 0.4940 0.6407 0.8005
No log 3.24 324 0.6600 0.4930 0.6600 0.8124
No log 3.26 326 0.6784 0.5295 0.6784 0.8236
No log 3.2800 328 0.5590 0.5131 0.5590 0.7477
No log 3.3 330 0.4872 0.3763 0.4872 0.6980
No log 3.32 332 0.4901 0.4878 0.4901 0.7001
No log 3.34 334 0.4911 0.5213 0.4911 0.7008
No log 3.36 336 0.5213 0.5081 0.5213 0.7220
No log 3.38 338 0.5777 0.5562 0.5777 0.7601
No log 3.4 340 0.5336 0.5140 0.5336 0.7305
No log 3.42 342 0.5038 0.4642 0.5038 0.7098
No log 3.44 344 0.5023 0.4639 0.5023 0.7087
No log 3.46 346 0.5104 0.4314 0.5104 0.7144
No log 3.48 348 0.5118 0.4402 0.5118 0.7154
No log 3.5 350 0.5259 0.3919 0.5259 0.7252
No log 3.52 352 0.6353 0.4815 0.6353 0.7970
No log 3.54 354 0.7651 0.4541 0.7651 0.8747
No log 3.56 356 0.7617 0.4347 0.7617 0.8727
No log 3.58 358 0.6819 0.4959 0.6819 0.8258
No log 3.6 360 0.6344 0.5113 0.6344 0.7965
No log 3.62 362 0.5907 0.5131 0.5907 0.7686
No log 3.64 364 0.5699 0.5219 0.5699 0.7549
No log 3.66 366 0.5659 0.5219 0.5659 0.7523
No log 3.68 368 0.5448 0.5349 0.5448 0.7381
No log 3.7 370 0.5420 0.5111 0.5420 0.7362
No log 3.7200 372 0.5440 0.5111 0.5440 0.7376
No log 3.74 374 0.5328 0.4459 0.5328 0.7299
No log 3.76 376 0.5608 0.4315 0.5608 0.7489
No log 3.7800 378 0.6015 0.4606 0.6015 0.7756
No log 3.8 380 0.6127 0.4819 0.6127 0.7827
No log 3.82 382 0.6607 0.5027 0.6607 0.8128
No log 3.84 384 0.6560 0.5027 0.6560 0.8099
No log 3.86 386 0.6135 0.4892 0.6135 0.7833
No log 3.88 388 0.5263 0.4437 0.5263 0.7255
No log 3.9 390 0.5103 0.4402 0.5103 0.7143
No log 3.92 392 0.5171 0.5028 0.5171 0.7191
No log 3.94 394 0.5051 0.4229 0.5051 0.7107
No log 3.96 396 0.5341 0.4875 0.5341 0.7308
No log 3.98 398 0.6487 0.5325 0.6487 0.8054
No log 4.0 400 0.6871 0.5295 0.6871 0.8289
No log 4.02 402 0.6139 0.4978 0.6139 0.7835
No log 4.04 404 0.5173 0.4795 0.5173 0.7192
No log 4.06 406 0.5236 0.3988 0.5236 0.7236
No log 4.08 408 0.5491 0.5065 0.5491 0.7410
No log 4.1 410 0.5365 0.5110 0.5365 0.7324
No log 4.12 412 0.5256 0.4206 0.5256 0.7250
No log 4.14 414 0.6206 0.5326 0.6206 0.7878
No log 4.16 416 0.7311 0.4601 0.7311 0.8550
No log 4.18 418 0.7374 0.4462 0.7374 0.8587
No log 4.2 420 0.6355 0.4815 0.6355 0.7972
No log 4.22 422 0.5739 0.3840 0.5739 0.7576
No log 4.24 424 0.5675 0.3919 0.5675 0.7533
No log 4.26 426 0.5719 0.3840 0.5719 0.7563
No log 4.28 428 0.6036 0.4835 0.6036 0.7769
No log 4.3 430 0.6301 0.4979 0.6301 0.7938
No log 4.32 432 0.5941 0.4270 0.5941 0.7708
No log 4.34 434 0.5668 0.3688 0.5668 0.7529
No log 4.36 436 0.5820 0.4905 0.5820 0.7629
No log 4.38 438 0.5967 0.4815 0.5967 0.7725
No log 4.4 440 0.6562 0.5183 0.6562 0.8101
No log 4.42 442 0.6859 0.4404 0.6859 0.8282
No log 4.44 444 0.6350 0.4959 0.6350 0.7969
No log 4.46 446 0.5644 0.3688 0.5644 0.7512
No log 4.48 448 0.5244 0.5344 0.5244 0.7242
No log 4.5 450 0.5298 0.5160 0.5298 0.7279
No log 4.52 452 0.5242 0.4809 0.5242 0.7240
No log 4.54 454 0.5335 0.4569 0.5335 0.7304
No log 4.5600 456 0.6017 0.4574 0.6017 0.7757
No log 4.58 458 0.6588 0.5665 0.6588 0.8117
No log 4.6 460 0.6455 0.5665 0.6455 0.8035
No log 4.62 462 0.5768 0.4502 0.5768 0.7595
No log 4.64 464 0.5252 0.4027 0.5252 0.7247
No log 4.66 466 0.5270 0.4857 0.5270 0.7260
No log 4.68 468 0.5328 0.3728 0.5328 0.7299
No log 4.7 470 0.5522 0.3701 0.5522 0.7431
No log 4.72 472 0.5794 0.4592 0.5794 0.7612
No log 4.74 474 0.5950 0.5233 0.5950 0.7714
No log 4.76 476 0.6270 0.5048 0.6270 0.7918
No log 4.78 478 0.6646 0.5249 0.6646 0.8152
No log 4.8 480 0.6558 0.5249 0.6558 0.8098
No log 4.82 482 0.6476 0.5249 0.6476 0.8047
No log 4.84 484 0.6499 0.5548 0.6499 0.8062
No log 4.86 486 0.6366 0.5418 0.6366 0.7979
No log 4.88 488 0.6159 0.5233 0.6159 0.7848
No log 4.9 490 0.5684 0.3976 0.5684 0.7540
No log 4.92 492 0.5504 0.3701 0.5504 0.7419
No log 4.9400 494 0.5266 0.4768 0.5266 0.7257
No log 4.96 496 0.5216 0.5133 0.5216 0.7222
No log 4.98 498 0.5291 0.4719 0.5291 0.7274
0.3244 5.0 500 0.5346 0.5125 0.5346 0.7312
0.3244 5.02 502 0.5547 0.4845 0.5547 0.7448
0.3244 5.04 504 0.5754 0.5101 0.5754 0.7585
0.3244 5.06 506 0.5656 0.4335 0.5656 0.7520
0.3244 5.08 508 0.5544 0.4134 0.5544 0.7446
0.3244 5.1 510 0.5460 0.4774 0.5460 0.7389

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k20_task7_organization

Finetuned
(4019)
this model