ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k12_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5891
  • Qwk: 0.4856
  • Mse: 0.5891
  • Rmse: 0.7675

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0328 2 2.4902 -0.0924 2.4902 1.5780
No log 0.0656 4 1.1697 0.1265 1.1697 1.0815
No log 0.0984 6 0.9190 0.0955 0.9190 0.9587
No log 0.1311 8 0.8454 0.0778 0.8454 0.9195
No log 0.1639 10 0.9588 0.0327 0.9588 0.9792
No log 0.1967 12 1.1822 0.0692 1.1822 1.0873
No log 0.2295 14 1.2553 0.0576 1.2553 1.1204
No log 0.2623 16 1.1568 0.0896 1.1568 1.0756
No log 0.2951 18 1.1175 0.1208 1.1175 1.0571
No log 0.3279 20 1.0352 0.0719 1.0352 1.0175
No log 0.3607 22 0.9616 0.1285 0.9616 0.9806
No log 0.3934 24 0.9525 -0.0020 0.9525 0.9760
No log 0.4262 26 0.9570 0.1089 0.9570 0.9783
No log 0.4590 28 0.8598 0.2019 0.8598 0.9273
No log 0.4918 30 0.8008 0.1479 0.8008 0.8949
No log 0.5246 32 0.8052 0.1867 0.8052 0.8973
No log 0.5574 34 0.8922 0.2979 0.8922 0.9445
No log 0.5902 36 1.0731 0.0912 1.0731 1.0359
No log 0.6230 38 1.0655 0.0569 1.0655 1.0322
No log 0.6557 40 1.0214 0.1155 1.0214 1.0106
No log 0.6885 42 0.8687 0.1884 0.8687 0.9321
No log 0.7213 44 0.8696 0.2604 0.8696 0.9325
No log 0.7541 46 0.8819 0.1495 0.8819 0.9391
No log 0.7869 48 1.0503 0.0844 1.0503 1.0248
No log 0.8197 50 1.4484 0.0338 1.4484 1.2035
No log 0.8525 52 1.6178 -0.0064 1.6178 1.2719
No log 0.8852 54 1.5125 -0.0295 1.5125 1.2298
No log 0.9180 56 1.2326 0.0736 1.2326 1.1102
No log 0.9508 58 0.9777 0.0856 0.9777 0.9888
No log 0.9836 60 0.8506 0.0265 0.8506 0.9223
No log 1.0164 62 0.8394 0.1648 0.8394 0.9162
No log 1.0492 64 0.8565 0.2467 0.8565 0.9255
No log 1.0820 66 0.8502 0.2227 0.8502 0.9221
No log 1.1148 68 0.8507 0.0154 0.8507 0.9223
No log 1.1475 70 0.8663 -0.0121 0.8663 0.9307
No log 1.1803 72 0.8656 0.0313 0.8656 0.9304
No log 1.2131 74 0.8620 -0.0026 0.8620 0.9284
No log 1.2459 76 0.8471 -0.0192 0.8471 0.9204
No log 1.2787 78 0.8671 0.2227 0.8671 0.9312
No log 1.3115 80 0.8559 0.0860 0.8559 0.9251
No log 1.3443 82 0.8519 0.0662 0.8519 0.9230
No log 1.3770 84 0.8458 0.0662 0.8458 0.9197
No log 1.4098 86 0.8240 0.1463 0.8240 0.9078
No log 1.4426 88 0.8378 0.1647 0.8378 0.9153
No log 1.4754 90 0.8512 0.2015 0.8512 0.9226
No log 1.5082 92 0.8797 0.1889 0.8797 0.9379
No log 1.5410 94 0.9150 0.2550 0.9150 0.9566
No log 1.5738 96 1.0033 0.1506 1.0033 1.0017
No log 1.6066 98 1.0439 0.1241 1.0439 1.0217
No log 1.6393 100 1.0377 0.1206 1.0377 1.0187
No log 1.6721 102 0.9512 0.2047 0.9512 0.9753
No log 1.7049 104 0.9200 0.2600 0.9200 0.9592
No log 1.7377 106 0.8501 0.3523 0.8501 0.9220
No log 1.7705 108 0.7985 0.3536 0.7985 0.8936
No log 1.8033 110 0.8327 0.3613 0.8327 0.9125
No log 1.8361 112 0.8063 0.2549 0.8063 0.8979
No log 1.8689 114 0.7273 0.3754 0.7273 0.8528
No log 1.9016 116 0.7235 0.3390 0.7235 0.8506
No log 1.9344 118 0.7376 0.3704 0.7376 0.8588
No log 1.9672 120 0.7652 0.3739 0.7652 0.8748
No log 2.0 122 0.7867 0.3378 0.7867 0.8870
No log 2.0328 124 0.7765 0.3941 0.7765 0.8812
No log 2.0656 126 0.7421 0.3785 0.7421 0.8614
No log 2.0984 128 0.7296 0.3930 0.7296 0.8542
No log 2.1311 130 0.7261 0.3542 0.7261 0.8521
No log 2.1639 132 0.7428 0.3829 0.7428 0.8619
No log 2.1967 134 0.7346 0.3936 0.7346 0.8571
No log 2.2295 136 0.7794 0.3392 0.7794 0.8828
No log 2.2623 138 0.7396 0.3683 0.7396 0.8600
No log 2.2951 140 0.7307 0.2884 0.7307 0.8548
No log 2.3279 142 0.7785 0.3988 0.7785 0.8823
No log 2.3607 144 0.8766 0.3559 0.8766 0.9363
No log 2.3934 146 0.8526 0.4396 0.8526 0.9233
No log 2.4262 148 0.8250 0.3896 0.8250 0.9083
No log 2.4590 150 0.9111 0.2382 0.9111 0.9545
No log 2.4918 152 1.0682 0.2125 1.0682 1.0336
No log 2.5246 154 1.0037 0.2725 1.0037 1.0018
No log 2.5574 156 0.8467 0.4140 0.8467 0.9201
No log 2.5902 158 0.8014 0.4165 0.8014 0.8952
No log 2.6230 160 0.7790 0.4051 0.7790 0.8826
No log 2.6557 162 0.7818 0.4122 0.7818 0.8842
No log 2.6885 164 0.7562 0.3937 0.7562 0.8696
No log 2.7213 166 0.7208 0.3590 0.7208 0.8490
No log 2.7541 168 0.7070 0.3474 0.7070 0.8408
No log 2.7869 170 0.7184 0.3950 0.7184 0.8476
No log 2.8197 172 0.7237 0.4006 0.7237 0.8507
No log 2.8525 174 0.7313 0.4006 0.7313 0.8552
No log 2.8852 176 0.7491 0.4190 0.7491 0.8655
No log 2.9180 178 0.7530 0.4 0.7530 0.8678
No log 2.9508 180 0.7624 0.2790 0.7624 0.8732
No log 2.9836 182 0.7472 0.3259 0.7472 0.8644
No log 3.0164 184 0.7450 0.3622 0.7450 0.8631
No log 3.0492 186 0.7529 0.3178 0.7529 0.8677
No log 3.0820 188 0.7530 0.3768 0.7530 0.8678
No log 3.1148 190 0.7403 0.3995 0.7403 0.8604
No log 3.1475 192 0.7237 0.4377 0.7237 0.8507
No log 3.1803 194 0.6817 0.4484 0.6817 0.8256
No log 3.2131 196 0.6923 0.4044 0.6923 0.8321
No log 3.2459 198 0.6984 0.3899 0.6984 0.8357
No log 3.2787 200 0.7266 0.4131 0.7266 0.8524
No log 3.3115 202 0.7679 0.4076 0.7679 0.8763
No log 3.3443 204 0.8034 0.3579 0.8034 0.8963
No log 3.3770 206 0.7668 0.3913 0.7668 0.8757
No log 3.4098 208 0.6841 0.4461 0.6841 0.8271
No log 3.4426 210 0.6649 0.4364 0.6649 0.8154
No log 3.4754 212 0.6682 0.4364 0.6682 0.8175
No log 3.5082 214 0.6804 0.4517 0.6804 0.8249
No log 3.5410 216 0.6831 0.4619 0.6831 0.8265
No log 3.5738 218 0.6922 0.4256 0.6922 0.8320
No log 3.6066 220 0.7142 0.3856 0.7142 0.8451
No log 3.6393 222 0.7414 0.3814 0.7414 0.8610
No log 3.6721 224 0.7548 0.4167 0.7548 0.8688
No log 3.7049 226 0.7467 0.3959 0.7467 0.8641
No log 3.7377 228 0.7122 0.3811 0.7122 0.8439
No log 3.7705 230 0.6875 0.4991 0.6875 0.8292
No log 3.8033 232 0.6649 0.3887 0.6649 0.8154
No log 3.8361 234 0.6579 0.3127 0.6579 0.8111
No log 3.8689 236 0.6335 0.3622 0.6335 0.7959
No log 3.9016 238 0.6206 0.3995 0.6206 0.7878
No log 3.9344 240 0.6729 0.3716 0.6729 0.8203
No log 3.9672 242 0.6821 0.3789 0.6821 0.8259
No log 4.0 244 0.6938 0.4461 0.6938 0.8330
No log 4.0328 246 0.7173 0.4736 0.7173 0.8470
No log 4.0656 248 0.7221 0.4032 0.7221 0.8497
No log 4.0984 250 0.7178 0.3935 0.7178 0.8472
No log 4.1311 252 0.6880 0.3979 0.6880 0.8294
No log 4.1639 254 0.6438 0.5104 0.6438 0.8024
No log 4.1967 256 0.6541 0.5235 0.6541 0.8087
No log 4.2295 258 0.6697 0.4562 0.6697 0.8183
No log 4.2623 260 0.6598 0.5037 0.6598 0.8123
No log 4.2951 262 0.6605 0.5087 0.6605 0.8127
No log 4.3279 264 0.7092 0.3829 0.7092 0.8422
No log 4.3607 266 0.6655 0.4261 0.6655 0.8158
No log 4.3934 268 0.6256 0.5110 0.6256 0.7910
No log 4.4262 270 0.5963 0.5565 0.5963 0.7722
No log 4.4590 272 0.5886 0.6344 0.5886 0.7672
No log 4.4918 274 0.6178 0.4518 0.6178 0.7860
No log 4.5246 276 0.6307 0.4747 0.6307 0.7941
No log 4.5574 278 0.5806 0.5826 0.5806 0.7619
No log 4.5902 280 0.5689 0.5584 0.5689 0.7543
No log 4.6230 282 0.5862 0.5586 0.5862 0.7656
No log 4.6557 284 0.6123 0.5736 0.6123 0.7825
No log 4.6885 286 0.6026 0.5966 0.6026 0.7762
No log 4.7213 288 0.5986 0.5767 0.5986 0.7737
No log 4.7541 290 0.6368 0.5687 0.6368 0.7980
No log 4.7869 292 0.6227 0.5420 0.6227 0.7891
No log 4.8197 294 0.5870 0.6059 0.5870 0.7662
No log 4.8525 296 0.5815 0.6161 0.5815 0.7626
No log 4.8852 298 0.5860 0.5728 0.5860 0.7655
No log 4.9180 300 0.6392 0.5421 0.6392 0.7995
No log 4.9508 302 0.6113 0.5205 0.6113 0.7818
No log 4.9836 304 0.5658 0.5507 0.5658 0.7522
No log 5.0164 306 0.5996 0.5501 0.5996 0.7743
No log 5.0492 308 0.6001 0.5501 0.6001 0.7746
No log 5.0820 310 0.5434 0.5440 0.5434 0.7371
No log 5.1148 312 0.6160 0.5849 0.6160 0.7848
No log 5.1475 314 0.6867 0.5281 0.6867 0.8287
No log 5.1803 316 0.6202 0.5328 0.6202 0.7875
No log 5.2131 318 0.5331 0.5702 0.5331 0.7302
No log 5.2459 320 0.5318 0.5398 0.5318 0.7293
No log 5.2787 322 0.5432 0.5718 0.5432 0.7370
No log 5.3115 324 0.6199 0.5003 0.6199 0.7873
No log 5.3443 326 0.5754 0.5922 0.5754 0.7586
No log 5.3770 328 0.5133 0.5826 0.5133 0.7164
No log 5.4098 330 0.5112 0.5234 0.5112 0.7150
No log 5.4426 332 0.5233 0.6034 0.5233 0.7234
No log 5.4754 334 0.6026 0.5190 0.6026 0.7762
No log 5.5082 336 0.6505 0.5205 0.6505 0.8066
No log 5.5410 338 0.6270 0.5312 0.6270 0.7918
No log 5.5738 340 0.6290 0.5355 0.6290 0.7931
No log 5.6066 342 0.6226 0.5180 0.6226 0.7891
No log 5.6393 344 0.6490 0.4946 0.6490 0.8056
No log 5.6721 346 0.7425 0.4946 0.7425 0.8617
No log 5.7049 348 0.8865 0.4134 0.8865 0.9415
No log 5.7377 350 0.9062 0.4092 0.9062 0.9519
No log 5.7705 352 0.7650 0.4703 0.7650 0.8747
No log 5.8033 354 0.5902 0.5313 0.5902 0.7682
No log 5.8361 356 0.5394 0.5420 0.5394 0.7345
No log 5.8689 358 0.5459 0.6233 0.5459 0.7388
No log 5.9016 360 0.5926 0.4898 0.5926 0.7698
No log 5.9344 362 0.6716 0.4946 0.6716 0.8195
No log 5.9672 364 0.7551 0.4562 0.7551 0.8689
No log 6.0 366 0.8228 0.3847 0.8228 0.9071
No log 6.0328 368 0.8066 0.3847 0.8066 0.8981
No log 6.0656 370 0.7227 0.4756 0.7227 0.8501
No log 6.0984 372 0.6287 0.4761 0.6287 0.7929
No log 6.1311 374 0.6030 0.4833 0.6030 0.7765
No log 6.1639 376 0.6078 0.4783 0.6078 0.7796
No log 6.1967 378 0.5938 0.4981 0.5938 0.7706
No log 6.2295 380 0.6053 0.4800 0.6053 0.7780
No log 6.2623 382 0.5790 0.5267 0.5790 0.7609
No log 6.2951 384 0.5753 0.5250 0.5753 0.7585
No log 6.3279 386 0.5728 0.5037 0.5728 0.7568
No log 6.3607 388 0.5811 0.5037 0.5811 0.7623
No log 6.3934 390 0.6219 0.4864 0.6219 0.7886
No log 6.4262 392 0.7206 0.4860 0.7206 0.8489
No log 6.4590 394 0.7026 0.4521 0.7026 0.8382
No log 6.4918 396 0.5809 0.5647 0.5809 0.7621
No log 6.5246 398 0.6073 0.4783 0.6073 0.7793
No log 6.5574 400 0.6538 0.5013 0.6538 0.8086
No log 6.5902 402 0.5950 0.4997 0.5950 0.7714
No log 6.6230 404 0.5649 0.6078 0.5649 0.7516
No log 6.6557 406 0.5763 0.5947 0.5763 0.7592
No log 6.6885 408 0.5599 0.5476 0.5599 0.7483
No log 6.7213 410 0.5756 0.4918 0.5756 0.7587
No log 6.7541 412 0.6884 0.4580 0.6884 0.8297
No log 6.7869 414 0.8219 0.4250 0.8219 0.9066
No log 6.8197 416 0.8650 0.3652 0.8650 0.9301
No log 6.8525 418 0.7458 0.4987 0.7458 0.8636
No log 6.8852 420 0.6326 0.5679 0.6326 0.7954
No log 6.9180 422 0.6981 0.5186 0.6981 0.8355
No log 6.9508 424 0.7431 0.4698 0.7431 0.8620
No log 6.9836 426 0.6644 0.5555 0.6644 0.8151
No log 7.0164 428 0.5999 0.5334 0.5999 0.7745
No log 7.0492 430 0.6295 0.5392 0.6295 0.7934
No log 7.0820 432 0.7937 0.4087 0.7937 0.8909
No log 7.1148 434 0.9018 0.3066 0.9018 0.9496
No log 7.1475 436 0.9169 0.3233 0.9169 0.9575
No log 7.1803 438 0.8080 0.3890 0.8080 0.8989
No log 7.2131 440 0.7276 0.5185 0.7276 0.8530
No log 7.2459 442 0.6817 0.5247 0.6817 0.8256
No log 7.2787 444 0.6517 0.5274 0.6517 0.8073
No log 7.3115 446 0.6322 0.5274 0.6322 0.7951
No log 7.3443 448 0.6209 0.5243 0.6209 0.7880
No log 7.3770 450 0.6725 0.4971 0.6725 0.8201
No log 7.4098 452 0.7799 0.4186 0.7799 0.8831
No log 7.4426 454 0.8270 0.4159 0.8270 0.9094
No log 7.4754 456 0.8086 0.4098 0.8086 0.8992
No log 7.5082 458 0.7887 0.4098 0.7887 0.8881
No log 7.5410 460 0.7181 0.4598 0.7181 0.8474
No log 7.5738 462 0.6632 0.4747 0.6632 0.8144
No log 7.6066 464 0.6260 0.4711 0.6260 0.7912
No log 7.6393 466 0.6077 0.4727 0.6077 0.7796
No log 7.6721 468 0.5954 0.4692 0.5954 0.7716
No log 7.7049 470 0.6316 0.4882 0.6316 0.7947
No log 7.7377 472 0.6822 0.4344 0.6822 0.8259
No log 7.7705 474 0.7153 0.4427 0.7153 0.8458
No log 7.8033 476 0.6673 0.4535 0.6673 0.8169
No log 7.8361 478 0.5830 0.5941 0.5830 0.7636
No log 7.8689 480 0.5590 0.6553 0.5590 0.7476
No log 7.9016 482 0.5719 0.6256 0.5719 0.7562
No log 7.9344 484 0.5958 0.5524 0.5958 0.7719
No log 7.9672 486 0.6364 0.5251 0.6364 0.7977
No log 8.0 488 0.6562 0.5061 0.6562 0.8101
No log 8.0328 490 0.6597 0.5061 0.6597 0.8122
No log 8.0656 492 0.5925 0.5061 0.5925 0.7698
No log 8.0984 494 0.5488 0.5877 0.5488 0.7408
No log 8.1311 496 0.5478 0.4866 0.5478 0.7401
No log 8.1639 498 0.5610 0.4418 0.5610 0.7490
0.3558 8.1967 500 0.5716 0.5463 0.5716 0.7560
0.3558 8.2295 502 0.5989 0.4935 0.5989 0.7739
0.3558 8.2623 504 0.6117 0.5090 0.6117 0.7821
0.3558 8.2951 506 0.5893 0.5087 0.5893 0.7676
0.3558 8.3279 508 0.5853 0.4973 0.5853 0.7650
0.3558 8.3607 510 0.5891 0.4856 0.5891 0.7675

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k12_task7_organization

Finetuned
(4019)
this model