ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k7_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5981
  • Qwk: 0.4123
  • Mse: 0.5981
  • Rmse: 0.7734

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1 2 2.5233 -0.0924 2.5233 1.5885
No log 0.2 4 1.3088 -0.0141 1.3088 1.1440
No log 0.3 6 0.8683 -0.0841 0.8683 0.9318
No log 0.4 8 0.7302 0.1232 0.7302 0.8545
No log 0.5 10 0.7162 0.1321 0.7162 0.8463
No log 0.6 12 0.6981 0.0851 0.6981 0.8355
No log 0.7 14 0.7524 0.2558 0.7524 0.8674
No log 0.8 16 0.8060 0.3173 0.8060 0.8978
No log 0.9 18 0.8405 0.2841 0.8405 0.9168
No log 1.0 20 0.7620 0.1372 0.7620 0.8729
No log 1.1 22 0.7405 -0.0500 0.7405 0.8605
No log 1.2 24 0.7865 0.1313 0.7865 0.8869
No log 1.3 26 0.7876 0.3099 0.7876 0.8875
No log 1.4 28 0.7940 0.1550 0.7940 0.8911
No log 1.5 30 0.7659 0.1007 0.7659 0.8751
No log 1.6 32 0.7374 0.1508 0.7374 0.8587
No log 1.7 34 0.6885 0.0717 0.6885 0.8298
No log 1.8 36 0.6536 0.1942 0.6536 0.8084
No log 1.9 38 0.6287 0.3019 0.6287 0.7929
No log 2.0 40 0.6474 0.2522 0.6474 0.8046
No log 2.1 42 0.6845 0.2464 0.6845 0.8273
No log 2.2 44 0.6942 0.2526 0.6942 0.8332
No log 2.3 46 0.6201 0.1903 0.6201 0.7875
No log 2.4 48 0.7380 0.3699 0.7380 0.8591
No log 2.5 50 0.8357 0.3409 0.8357 0.9142
No log 2.6 52 0.7688 0.2574 0.7688 0.8768
No log 2.7 54 0.6042 0.3274 0.6042 0.7773
No log 2.8 56 0.5931 0.3151 0.5931 0.7701
No log 2.9 58 0.5894 0.3499 0.5894 0.7677
No log 3.0 60 0.5806 0.3105 0.5806 0.7620
No log 3.1 62 0.5450 0.3151 0.5450 0.7383
No log 3.2 64 0.5750 0.4330 0.5750 0.7583
No log 3.3 66 0.6647 0.3799 0.6647 0.8153
No log 3.4 68 0.6143 0.3843 0.6143 0.7837
No log 3.5 70 0.5592 0.4795 0.5592 0.7478
No log 3.6 72 0.5599 0.5056 0.5599 0.7482
No log 3.7 74 0.6522 0.3544 0.6522 0.8076
No log 3.8 76 0.5816 0.3813 0.5816 0.7626
No log 3.9 78 0.5553 0.3945 0.5553 0.7452
No log 4.0 80 0.5704 0.4044 0.5704 0.7552
No log 4.1 82 0.6438 0.3662 0.6438 0.8024
No log 4.2 84 0.6417 0.2843 0.6417 0.8011
No log 4.3 86 0.5697 0.4322 0.5697 0.7548
No log 4.4 88 0.5366 0.4908 0.5366 0.7326
No log 4.5 90 0.5236 0.5373 0.5236 0.7236
No log 4.6 92 0.4962 0.5283 0.4962 0.7044
No log 4.7 94 0.4920 0.4938 0.4920 0.7014
No log 4.8 96 0.5321 0.4315 0.5321 0.7294
No log 4.9 98 0.7984 0.4542 0.7984 0.8936
No log 5.0 100 0.8853 0.4305 0.8853 0.9409
No log 5.1 102 0.6890 0.4646 0.6890 0.8301
No log 5.2 104 0.6207 0.4371 0.6207 0.7878
No log 5.3 106 0.5951 0.4473 0.5951 0.7714
No log 5.4 108 0.5120 0.5815 0.5120 0.7155
No log 5.5 110 0.5112 0.6377 0.5112 0.7150
No log 5.6 112 0.5175 0.5177 0.5175 0.7194
No log 5.7 114 0.6228 0.4550 0.6228 0.7892
No log 5.8 116 0.7527 0.4667 0.7527 0.8676
No log 5.9 118 0.6668 0.4197 0.6668 0.8166
No log 6.0 120 0.5901 0.5015 0.5901 0.7682
No log 6.1 122 0.5246 0.4937 0.5246 0.7243
No log 6.2 124 0.5191 0.5289 0.5191 0.7205
No log 6.3 126 0.5479 0.5357 0.5479 0.7402
No log 6.4 128 0.6104 0.4933 0.6104 0.7813
No log 6.5 130 0.7107 0.5175 0.7107 0.8430
No log 6.6 132 0.5343 0.5024 0.5343 0.7309
No log 6.7 134 0.4603 0.5555 0.4603 0.6785
No log 6.8 136 0.4604 0.5555 0.4604 0.6785
No log 6.9 138 0.4608 0.5846 0.4608 0.6788
No log 7.0 140 0.4754 0.6377 0.4754 0.6895
No log 7.1 142 0.5108 0.6053 0.5108 0.7147
No log 7.2 144 0.5322 0.5332 0.5322 0.7296
No log 7.3 146 0.5176 0.6492 0.5176 0.7194
No log 7.4 148 0.5027 0.5593 0.5027 0.7090
No log 7.5 150 0.5056 0.5549 0.5056 0.7111
No log 7.6 152 0.5095 0.4934 0.5095 0.7138
No log 7.7 154 0.5177 0.4878 0.5177 0.7195
No log 7.8 156 0.5354 0.5307 0.5354 0.7317
No log 7.9 158 0.5406 0.5307 0.5406 0.7353
No log 8.0 160 0.5876 0.3737 0.5876 0.7665
No log 8.1 162 0.6668 0.4424 0.6668 0.8166
No log 8.2 164 0.6012 0.4212 0.6012 0.7754
No log 8.3 166 0.5714 0.4816 0.5714 0.7559
No log 8.4 168 0.5815 0.4829 0.5815 0.7625
No log 8.5 170 0.5954 0.4693 0.5954 0.7716
No log 8.6 172 0.5834 0.6242 0.5834 0.7638
No log 8.7 174 0.6367 0.4473 0.6367 0.7979
No log 8.8 176 0.7533 0.4158 0.7533 0.8679
No log 8.9 178 0.6984 0.4837 0.6984 0.8357
No log 9.0 180 0.6235 0.5115 0.6235 0.7896
No log 9.1 182 0.6010 0.5143 0.6010 0.7752
No log 9.2 184 0.5890 0.4866 0.5890 0.7675
No log 9.3 186 0.6275 0.4788 0.6275 0.7921
No log 9.4 188 0.7551 0.4070 0.7551 0.8690
No log 9.5 190 0.7553 0.3699 0.7553 0.8691
No log 9.6 192 0.6226 0.4642 0.6226 0.7891
No log 9.7 194 0.6061 0.5452 0.6061 0.7786
No log 9.8 196 0.6367 0.4598 0.6367 0.7979
No log 9.9 198 0.5888 0.5596 0.5888 0.7674
No log 10.0 200 0.5742 0.4816 0.5742 0.7578
No log 10.1 202 0.6097 0.3471 0.6097 0.7808
No log 10.2 204 0.6128 0.3545 0.6128 0.7828
No log 10.3 206 0.5837 0.4267 0.5837 0.7640
No log 10.4 208 0.5720 0.4929 0.5720 0.7563
No log 10.5 210 0.5737 0.4816 0.5737 0.7574
No log 10.6 212 0.6049 0.3737 0.6049 0.7778
No log 10.7 214 0.6508 0.4036 0.6508 0.8067
No log 10.8 216 0.7095 0.3869 0.7095 0.8423
No log 10.9 218 0.7160 0.3869 0.7160 0.8461
No log 11.0 220 0.6757 0.3891 0.6757 0.8220
No log 11.1 222 0.6250 0.3662 0.6250 0.7906
No log 11.2 224 0.5918 0.4212 0.5918 0.7693
No log 11.3 226 0.5606 0.4724 0.5606 0.7487
No log 11.4 228 0.5553 0.5357 0.5553 0.7452
No log 11.5 230 0.5789 0.4674 0.5789 0.7609
No log 11.6 232 0.5646 0.4298 0.5646 0.7514
No log 11.7 234 0.5649 0.4182 0.5649 0.7516
No log 11.8 236 0.6379 0.3996 0.6379 0.7987
No log 11.9 238 0.7265 0.4531 0.7265 0.8523
No log 12.0 240 0.7145 0.4531 0.7145 0.8453
No log 12.1 242 0.6517 0.4759 0.6517 0.8073
No log 12.2 244 0.5862 0.4836 0.5862 0.7656
No log 12.3 246 0.5761 0.4764 0.5761 0.7590
No log 12.4 248 0.5970 0.4764 0.5970 0.7726
No log 12.5 250 0.6742 0.4167 0.6742 0.8211
No log 12.6 252 0.7245 0.3822 0.7245 0.8512
No log 12.7 254 0.6558 0.4134 0.6558 0.8098
No log 12.8 256 0.5993 0.4211 0.5993 0.7742
No log 12.9 258 0.6316 0.5041 0.6316 0.7947
No log 13.0 260 0.6127 0.4581 0.6127 0.7827
No log 13.1 262 0.6026 0.4244 0.6026 0.7763
No log 13.2 264 0.6951 0.4059 0.6951 0.8337
No log 13.3 266 0.8889 0.4297 0.8889 0.9428
No log 13.4 268 0.8964 0.4076 0.8964 0.9468
No log 13.5 270 0.7148 0.4424 0.7148 0.8455
No log 13.6 272 0.5591 0.4044 0.5591 0.7477
No log 13.7 274 0.5349 0.4067 0.5349 0.7314
No log 13.8 276 0.5330 0.4067 0.5330 0.7301
No log 13.9 278 0.5383 0.4044 0.5383 0.7337
No log 14.0 280 0.5329 0.4044 0.5329 0.7300
No log 14.1 282 0.5275 0.4618 0.5275 0.7263
No log 14.2 284 0.5293 0.4618 0.5293 0.7275
No log 14.3 286 0.5069 0.4397 0.5069 0.7120
No log 14.4 288 0.5033 0.4555 0.5033 0.7094
No log 14.5 290 0.5130 0.5039 0.5130 0.7162
No log 14.6 292 0.5334 0.5580 0.5334 0.7303
No log 14.7 294 0.5445 0.6014 0.5445 0.7379
No log 14.8 296 0.5483 0.5612 0.5483 0.7405
No log 14.9 298 0.5928 0.4393 0.5928 0.7699
No log 15.0 300 0.5877 0.3966 0.5877 0.7666
No log 15.1 302 0.5663 0.3966 0.5663 0.7525
No log 15.2 304 0.5477 0.4397 0.5477 0.7401
No log 15.3 306 0.5392 0.4044 0.5392 0.7343
No log 15.4 308 0.5884 0.3688 0.5884 0.7671
No log 15.5 310 0.6609 0.3819 0.6609 0.8130
No log 15.6 312 0.6424 0.3819 0.6424 0.8015
No log 15.7 314 0.5899 0.3867 0.5899 0.7681
No log 15.8 316 0.5484 0.3713 0.5484 0.7405
No log 15.9 318 0.5462 0.4278 0.5462 0.7390
No log 16.0 320 0.5680 0.4356 0.5680 0.7536
No log 16.1 322 0.6611 0.4684 0.6611 0.8131
No log 16.2 324 0.7071 0.4521 0.7071 0.8409
No log 16.3 326 0.6279 0.4531 0.6279 0.7924
No log 16.4 328 0.5847 0.4845 0.5847 0.7647
No log 16.5 330 0.5594 0.5184 0.5594 0.7479
No log 16.6 332 0.5379 0.4985 0.5379 0.7334
No log 16.7 334 0.5357 0.4234 0.5357 0.7319
No log 16.8 336 0.5311 0.4044 0.5311 0.7287
No log 16.9 338 0.5370 0.3763 0.5370 0.7328
No log 17.0 340 0.5756 0.3723 0.5756 0.7587
No log 17.1 342 0.6442 0.4587 0.6442 0.8026
No log 17.2 344 0.6957 0.4512 0.6957 0.8341
No log 17.3 346 0.6676 0.4512 0.6676 0.8171
No log 17.4 348 0.5659 0.3723 0.5659 0.7523
No log 17.5 350 0.5088 0.4555 0.5088 0.7133
No log 17.6 352 0.4971 0.5248 0.4971 0.7050
No log 17.7 354 0.4940 0.5267 0.4940 0.7029
No log 17.8 356 0.4980 0.5003 0.4980 0.7057
No log 17.9 358 0.5081 0.3701 0.5081 0.7128
No log 18.0 360 0.5490 0.3789 0.5490 0.7409
No log 18.1 362 0.5699 0.3789 0.5699 0.7549
No log 18.2 364 0.5555 0.4190 0.5555 0.7453
No log 18.3 366 0.5671 0.4451 0.5671 0.7530
No log 18.4 368 0.6036 0.4602 0.6036 0.7769
No log 18.5 370 0.5924 0.4602 0.5924 0.7697
No log 18.6 372 0.5595 0.4212 0.5595 0.7480
No log 18.7 374 0.5439 0.4100 0.5439 0.7375
No log 18.8 376 0.5518 0.3814 0.5518 0.7428
No log 18.9 378 0.5650 0.3814 0.5650 0.7517
No log 19.0 380 0.5788 0.3518 0.5788 0.7608
No log 19.1 382 0.5960 0.3518 0.5960 0.7720
No log 19.2 384 0.5901 0.3518 0.5901 0.7682
No log 19.3 386 0.6286 0.3518 0.6286 0.7929
No log 19.4 388 0.6380 0.3789 0.6380 0.7988
No log 19.5 390 0.6335 0.3789 0.6335 0.7959
No log 19.6 392 0.6015 0.4352 0.6015 0.7755
No log 19.7 394 0.5710 0.3763 0.5710 0.7556
No log 19.8 396 0.5456 0.4067 0.5456 0.7386
No log 19.9 398 0.5479 0.4067 0.5479 0.7402
No log 20.0 400 0.5586 0.4100 0.5586 0.7474
No log 20.1 402 0.5553 0.4100 0.5553 0.7452
No log 20.2 404 0.5535 0.4100 0.5535 0.7440
No log 20.3 406 0.5429 0.4378 0.5429 0.7368
No log 20.4 408 0.5452 0.3474 0.5452 0.7384
No log 20.5 410 0.5381 0.3474 0.5381 0.7336
No log 20.6 412 0.5312 0.4637 0.5312 0.7288
No log 20.7 414 0.5414 0.4375 0.5414 0.7358
No log 20.8 416 0.5453 0.4704 0.5453 0.7384
No log 20.9 418 0.5450 0.5081 0.5450 0.7382
No log 21.0 420 0.5242 0.4562 0.5242 0.7240
No log 21.1 422 0.5114 0.4726 0.5114 0.7151
No log 21.2 424 0.5273 0.3976 0.5273 0.7262
No log 21.3 426 0.5597 0.4219 0.5597 0.7482
No log 21.4 428 0.5727 0.4219 0.5727 0.7568
No log 21.5 430 0.5760 0.3918 0.5760 0.7589
No log 21.6 432 0.5816 0.4350 0.5816 0.7626
No log 21.7 434 0.5792 0.4350 0.5792 0.7611
No log 21.8 436 0.5526 0.4190 0.5526 0.7434
No log 21.9 438 0.5587 0.5081 0.5587 0.7475
No log 22.0 440 0.5560 0.5081 0.5560 0.7457
No log 22.1 442 0.5683 0.4704 0.5683 0.7538
No log 22.2 444 0.5358 0.4704 0.5358 0.7320
No log 22.3 446 0.5265 0.4639 0.5265 0.7256
No log 22.4 448 0.5159 0.5131 0.5159 0.7183
No log 22.5 450 0.4932 0.5151 0.4932 0.7023
No log 22.6 452 0.4904 0.5151 0.4904 0.7003
No log 22.7 454 0.5129 0.4618 0.5129 0.7162
No log 22.8 456 0.5182 0.4618 0.5182 0.7198
No log 22.9 458 0.5206 0.4618 0.5206 0.7215
No log 23.0 460 0.5128 0.4618 0.5128 0.7161
No log 23.1 462 0.4911 0.4964 0.4911 0.7008
No log 23.2 464 0.4979 0.4618 0.4979 0.7056
No log 23.3 466 0.5053 0.4618 0.5053 0.7109
No log 23.4 468 0.5390 0.4618 0.5390 0.7342
No log 23.5 470 0.5911 0.4684 0.5911 0.7689
No log 23.6 472 0.6792 0.4930 0.6792 0.8241
No log 23.7 474 0.6723 0.5147 0.6723 0.8199
No log 23.8 476 0.5952 0.4836 0.5952 0.7715
No log 23.9 478 0.5215 0.5254 0.5215 0.7222
No log 24.0 480 0.4918 0.5107 0.4918 0.7013
No log 24.1 482 0.4949 0.5522 0.4949 0.7035
No log 24.2 484 0.5029 0.4358 0.5029 0.7092
No log 24.3 486 0.5200 0.4639 0.5200 0.7211
No log 24.4 488 0.5347 0.4375 0.5347 0.7313
No log 24.5 490 0.5538 0.4100 0.5538 0.7442
No log 24.6 492 0.5445 0.4639 0.5445 0.7379
No log 24.7 494 0.5162 0.5141 0.5162 0.7185
No log 24.8 496 0.4953 0.5584 0.4953 0.7038
No log 24.9 498 0.4939 0.5323 0.4939 0.7028
0.3004 25.0 500 0.5112 0.5708 0.5112 0.7150
0.3004 25.1 502 0.5604 0.4997 0.5604 0.7486
0.3004 25.2 504 0.5813 0.5140 0.5813 0.7625
0.3004 25.3 506 0.5304 0.4864 0.5304 0.7283
0.3004 25.4 508 0.4776 0.5003 0.4776 0.6911
0.3004 25.5 510 0.4632 0.5930 0.4632 0.6806
0.3004 25.6 512 0.4817 0.6414 0.4817 0.6941
0.3004 25.7 514 0.4878 0.6414 0.4878 0.6984
0.3004 25.8 516 0.4690 0.6154 0.4690 0.6848
0.3004 25.9 518 0.4689 0.6024 0.4689 0.6848
0.3004 26.0 520 0.5495 0.5323 0.5495 0.7413
0.3004 26.1 522 0.6728 0.5325 0.6728 0.8202
0.3004 26.2 524 0.6964 0.4917 0.6964 0.8345
0.3004 26.3 526 0.6651 0.4153 0.6651 0.8155
0.3004 26.4 528 0.5981 0.4123 0.5981 0.7734

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k7_task7_organization

Finetuned
(4023)
this model