ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k16_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6372
  • Qwk: 0.4282
  • Mse: 0.6372
  • Rmse: 0.7982

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0364 2 2.5788 -0.0262 2.5788 1.6059
No log 0.0727 4 1.2167 0.1262 1.2167 1.1031
No log 0.1091 6 0.9711 -0.0646 0.9711 0.9854
No log 0.1455 8 0.9854 -0.0023 0.9854 0.9927
No log 0.1818 10 0.8059 -0.0963 0.8059 0.8977
No log 0.2182 12 0.7344 0.0481 0.7344 0.8570
No log 0.2545 14 0.8561 0.2769 0.8561 0.9252
No log 0.2909 16 1.0006 0.2113 1.0006 1.0003
No log 0.3273 18 0.9503 0.2202 0.9503 0.9748
No log 0.3636 20 0.8344 0.3131 0.8344 0.9135
No log 0.4 22 0.7770 0.2156 0.7770 0.8815
No log 0.4364 24 0.6959 0.1321 0.6959 0.8342
No log 0.4727 26 0.6919 0.2080 0.6919 0.8318
No log 0.5091 28 0.7118 0.3719 0.7118 0.8437
No log 0.5455 30 0.7827 0.4209 0.7827 0.8847
No log 0.5818 32 0.7997 0.4209 0.7997 0.8943
No log 0.6182 34 0.8237 0.3993 0.8237 0.9076
No log 0.6545 36 0.9472 0.2354 0.9472 0.9732
No log 0.6909 38 0.8210 0.3553 0.8210 0.9061
No log 0.7273 40 0.7009 0.2979 0.7009 0.8372
No log 0.7636 42 0.7409 0.1539 0.7409 0.8608
No log 0.8 44 0.8503 0.2156 0.8503 0.9221
No log 0.8364 46 1.0590 0.3425 1.0590 1.0291
No log 0.8727 48 1.0693 0.2784 1.0693 1.0341
No log 0.9091 50 0.9131 0.2132 0.9131 0.9556
No log 0.9455 52 0.7500 0.0481 0.7500 0.8660
No log 0.9818 54 0.7062 0.0 0.7062 0.8403
No log 1.0182 56 0.7527 0.2718 0.7527 0.8676
No log 1.0545 58 0.7625 0.2930 0.7625 0.8732
No log 1.0909 60 0.7079 0.1922 0.7079 0.8414
No log 1.1273 62 0.7109 0.0 0.7109 0.8431
No log 1.1636 64 0.9790 0.3425 0.9790 0.9894
No log 1.2 66 1.1827 0.1702 1.1827 1.0875
No log 1.2364 68 1.0662 0.3247 1.0662 1.0326
No log 1.2727 70 0.7904 0.2703 0.7904 0.8890
No log 1.3091 72 0.6162 0.1287 0.6162 0.7850
No log 1.3455 74 0.6215 0.1287 0.6215 0.7884
No log 1.3818 76 0.6702 0.1321 0.6702 0.8186
No log 1.4182 78 0.6998 0.0481 0.6998 0.8366
No log 1.4545 80 0.6769 0.0 0.6769 0.8227
No log 1.4909 82 0.6509 0.1236 0.6509 0.8068
No log 1.5273 84 0.7122 0.4219 0.7122 0.8439
No log 1.5636 86 0.7040 0.4479 0.7040 0.8391
No log 1.6 88 0.6505 0.2471 0.6505 0.8065
No log 1.6364 90 0.6424 0.2405 0.6424 0.8015
No log 1.6727 92 0.7043 0.3125 0.7043 0.8392
No log 1.7091 94 0.7380 0.3177 0.7380 0.8591
No log 1.7455 96 0.6569 0.4448 0.6569 0.8105
No log 1.7818 98 0.6514 0.4614 0.6514 0.8071
No log 1.8182 100 0.6359 0.4536 0.6359 0.7974
No log 1.8545 102 0.5823 0.5009 0.5823 0.7631
No log 1.8909 104 0.5805 0.5095 0.5805 0.7619
No log 1.9273 106 0.6063 0.4486 0.6063 0.7787
No log 1.9636 108 0.7217 0.4327 0.7217 0.8495
No log 2.0 110 0.7659 0.4007 0.7659 0.8752
No log 2.0364 112 0.6692 0.4594 0.6692 0.8181
No log 2.0727 114 0.5625 0.4970 0.5625 0.7500
No log 2.1091 116 0.6160 0.3737 0.6160 0.7848
No log 2.1455 118 0.6776 0.2285 0.6777 0.8232
No log 2.1818 120 0.6153 0.3167 0.6153 0.7844
No log 2.2182 122 0.5386 0.3575 0.5386 0.7339
No log 2.2545 124 0.5335 0.4289 0.5335 0.7304
No log 2.2909 126 0.5257 0.4561 0.5257 0.7250
No log 2.3273 128 0.5262 0.4492 0.5262 0.7254
No log 2.3636 130 0.5448 0.5114 0.5448 0.7381
No log 2.4 132 0.5893 0.5026 0.5893 0.7677
No log 2.4364 134 0.6241 0.5407 0.6241 0.7900
No log 2.4727 136 0.5995 0.5795 0.5995 0.7742
No log 2.5091 138 0.6043 0.5300 0.6043 0.7774
No log 2.5455 140 0.6165 0.5165 0.6165 0.7852
No log 2.5818 142 0.6498 0.4725 0.6498 0.8061
No log 2.6182 144 0.6441 0.5469 0.6441 0.8025
No log 2.6545 146 0.5929 0.5167 0.5929 0.7700
No log 2.6909 148 0.6121 0.4819 0.6121 0.7824
No log 2.7273 150 0.6068 0.4819 0.6068 0.7790
No log 2.7636 152 0.5780 0.5222 0.5780 0.7602
No log 2.8 154 0.5852 0.5092 0.5852 0.7650
No log 2.8364 156 0.5843 0.5151 0.5843 0.7644
No log 2.8727 158 0.5805 0.5383 0.5805 0.7619
No log 2.9091 160 0.5797 0.5563 0.5797 0.7614
No log 2.9455 162 0.6439 0.4930 0.6439 0.8024
No log 2.9818 164 0.6992 0.4332 0.6992 0.8362
No log 3.0182 166 0.7820 0.3942 0.7820 0.8843
No log 3.0545 168 0.7364 0.3806 0.7364 0.8581
No log 3.0909 170 0.6087 0.4595 0.6087 0.7802
No log 3.1273 172 0.5547 0.5734 0.5547 0.7448
No log 3.1636 174 0.5811 0.5133 0.5811 0.7623
No log 3.2 176 0.5917 0.5533 0.5917 0.7692
No log 3.2364 178 0.5880 0.5549 0.5880 0.7668
No log 3.2727 180 0.6032 0.5476 0.6032 0.7766
No log 3.3091 182 0.6274 0.4783 0.6274 0.7921
No log 3.3455 184 0.6827 0.4224 0.6827 0.8262
No log 3.3818 186 0.6896 0.4344 0.6896 0.8304
No log 3.4182 188 0.6251 0.4817 0.6251 0.7906
No log 3.4545 190 0.6426 0.4627 0.6426 0.8016
No log 3.4909 192 0.7551 0.4369 0.7551 0.8690
No log 3.5273 194 0.7885 0.4152 0.7885 0.8880
No log 3.5636 196 0.6507 0.4627 0.6507 0.8066
No log 3.6 198 0.6163 0.4424 0.6163 0.7850
No log 3.6364 200 0.7351 0.4668 0.7351 0.8574
No log 3.6727 202 0.6980 0.4768 0.6980 0.8355
No log 3.7091 204 0.5842 0.5014 0.5842 0.7643
No log 3.7455 206 0.5925 0.5404 0.5925 0.7697
No log 3.7818 208 0.7560 0.4877 0.7560 0.8695
No log 3.8182 210 0.7497 0.4877 0.7497 0.8658
No log 3.8545 212 0.6130 0.4724 0.6130 0.7829
No log 3.8909 214 0.5854 0.4114 0.5854 0.7651
No log 3.9273 216 0.6056 0.3837 0.6056 0.7782
No log 3.9636 218 0.6220 0.3837 0.6220 0.7887
No log 4.0 220 0.6524 0.3990 0.6524 0.8077
No log 4.0364 222 0.6502 0.4073 0.6502 0.8063
No log 4.0727 224 0.6524 0.4322 0.6524 0.8077
No log 4.1091 226 0.6394 0.4735 0.6394 0.7996
No log 4.1455 228 0.6388 0.4024 0.6388 0.7993
No log 4.1818 230 0.7366 0.4332 0.7366 0.8582
No log 4.2182 232 0.7305 0.4287 0.7305 0.8547
No log 4.2545 234 0.6357 0.3988 0.6357 0.7973
No log 4.2909 236 0.5748 0.4082 0.5748 0.7581
No log 4.3273 238 0.6079 0.4234 0.6079 0.7797
No log 4.3636 240 0.6217 0.4473 0.6217 0.7885
No log 4.4 242 0.5952 0.4582 0.5952 0.7715
No log 4.4364 244 0.5809 0.3910 0.5809 0.7622
No log 4.4727 246 0.5833 0.5217 0.5833 0.7638
No log 4.5091 248 0.5741 0.5809 0.5741 0.7577
No log 4.5455 250 0.5622 0.6235 0.5622 0.7498
No log 4.5818 252 0.5654 0.5881 0.5654 0.7520
No log 4.6182 254 0.5541 0.5765 0.5541 0.7444
No log 4.6545 256 0.5725 0.5212 0.5725 0.7566
No log 4.6909 258 0.6531 0.4825 0.6531 0.8081
No log 4.7273 260 0.6205 0.5299 0.6205 0.7877
No log 4.7636 262 0.5472 0.4776 0.5472 0.7397
No log 4.8 264 0.5835 0.5015 0.5835 0.7639
No log 4.8364 266 0.6196 0.5149 0.6196 0.7871
No log 4.8727 268 0.5665 0.4227 0.5665 0.7527
No log 4.9091 270 0.5713 0.4795 0.5713 0.7559
No log 4.9455 272 0.5941 0.4752 0.5941 0.7708
No log 4.9818 274 0.6145 0.4636 0.6145 0.7839
No log 5.0182 276 0.6115 0.4980 0.6115 0.7820
No log 5.0545 278 0.5826 0.4137 0.5826 0.7633
No log 5.0909 280 0.5758 0.4823 0.5758 0.7588
No log 5.1273 282 0.5803 0.48 0.5803 0.7617
No log 5.1636 284 0.5601 0.5866 0.5601 0.7484
No log 5.2 286 0.5535 0.6052 0.5535 0.7440
No log 5.2364 288 0.5484 0.5769 0.5484 0.7405
No log 5.2727 290 0.5299 0.5739 0.5299 0.7280
No log 5.3091 292 0.5098 0.5222 0.5098 0.7140
No log 5.3455 294 0.5049 0.5517 0.5049 0.7106
No log 5.3818 296 0.5023 0.5604 0.5023 0.7088
No log 5.4182 298 0.5067 0.5753 0.5067 0.7118
No log 5.4545 300 0.5061 0.5725 0.5061 0.7114
No log 5.4909 302 0.5150 0.5904 0.5150 0.7177
No log 5.5273 304 0.5426 0.5407 0.5426 0.7366
No log 5.5636 306 0.5501 0.5136 0.5501 0.7417
No log 5.6 308 0.6445 0.5313 0.6445 0.8028
No log 5.6364 310 0.6617 0.4003 0.6617 0.8134
No log 5.6727 312 0.5643 0.4635 0.5643 0.7512
No log 5.7091 314 0.5528 0.4575 0.5528 0.7435
No log 5.7455 316 0.5702 0.4517 0.5702 0.7551
No log 5.7818 318 0.6753 0.3957 0.6753 0.8218
No log 5.8182 320 0.8148 0.3886 0.8148 0.9027
No log 5.8545 322 0.7778 0.4333 0.7778 0.8819
No log 5.8909 324 0.6221 0.5078 0.6221 0.7887
No log 5.9273 326 0.5840 0.4051 0.5840 0.7642
No log 5.9636 328 0.5997 0.4428 0.5997 0.7744
No log 6.0 330 0.5921 0.4137 0.5921 0.7695
No log 6.0364 332 0.6172 0.4149 0.6172 0.7856
No log 6.0727 334 0.6302 0.4149 0.6302 0.7938
No log 6.1091 336 0.6037 0.3915 0.6037 0.7770
No log 6.1455 338 0.6059 0.3572 0.6059 0.7784
No log 6.1818 340 0.6510 0.4502 0.6510 0.8069
No log 6.2182 342 0.6481 0.4437 0.6481 0.8051
No log 6.2545 344 0.6106 0.3865 0.6106 0.7814
No log 6.2909 346 0.6051 0.3865 0.6051 0.7778
No log 6.3273 348 0.5914 0.4455 0.5914 0.7690
No log 6.3636 350 0.5946 0.4287 0.5946 0.7711
No log 6.4 352 0.6456 0.4518 0.6456 0.8035
No log 6.4364 354 0.6511 0.4302 0.6511 0.8069
No log 6.4727 356 0.6141 0.4370 0.6141 0.7836
No log 6.5091 358 0.5981 0.4763 0.5981 0.7733
No log 6.5455 360 0.6094 0.4493 0.6094 0.7807
No log 6.5818 362 0.6074 0.4763 0.6074 0.7794
No log 6.6182 364 0.6107 0.4051 0.6107 0.7815
No log 6.6545 366 0.6568 0.4713 0.6568 0.8104
No log 6.6909 368 0.6606 0.4150 0.6606 0.8128
No log 6.7273 370 0.6122 0.4402 0.6122 0.7824
No log 6.7636 372 0.6062 0.4871 0.6062 0.7786
No log 6.8 374 0.6131 0.4871 0.6131 0.7830
No log 6.8364 376 0.6156 0.5075 0.6156 0.7846
No log 6.8727 378 0.6345 0.4983 0.6345 0.7966
No log 6.9091 380 0.7059 0.5124 0.7059 0.8402
No log 6.9455 382 0.6824 0.5124 0.6824 0.8261
No log 6.9818 384 0.6207 0.5171 0.6207 0.7879
No log 7.0182 386 0.5923 0.5259 0.5923 0.7696
No log 7.0545 388 0.5838 0.5259 0.5838 0.7641
No log 7.0909 390 0.5849 0.4886 0.5849 0.7648
No log 7.1273 392 0.6437 0.5055 0.6437 0.8023
No log 7.1636 394 0.6543 0.5240 0.6543 0.8089
No log 7.2 396 0.6237 0.5419 0.6237 0.7897
No log 7.2364 398 0.5927 0.5629 0.5927 0.7699
No log 7.2727 400 0.5988 0.5285 0.5988 0.7738
No log 7.3091 402 0.6080 0.5387 0.6080 0.7798
No log 7.3455 404 0.7100 0.4719 0.7100 0.8426
No log 7.3818 406 0.8059 0.3654 0.8059 0.8977
No log 7.4182 408 0.7894 0.3690 0.7894 0.8885
No log 7.4545 410 0.6731 0.4189 0.6731 0.8205
No log 7.4909 412 0.5985 0.4776 0.5985 0.7736
No log 7.5273 414 0.5686 0.4569 0.5686 0.7540
No log 7.5636 416 0.5471 0.4463 0.5471 0.7397
No log 7.6 418 0.5402 0.4402 0.5402 0.7350
No log 7.6364 420 0.5422 0.5248 0.5422 0.7363
No log 7.6727 422 0.5499 0.5234 0.5499 0.7416
No log 7.7091 424 0.5397 0.4992 0.5397 0.7346
No log 7.7455 426 0.5239 0.5640 0.5239 0.7238
No log 7.7818 428 0.5266 0.5648 0.5266 0.7257
No log 7.8182 430 0.5489 0.6108 0.5489 0.7409
No log 7.8545 432 0.6013 0.6072 0.6013 0.7754
No log 7.8909 434 0.6169 0.5464 0.6169 0.7854
No log 7.9273 436 0.5403 0.5455 0.5403 0.7351
No log 7.9636 438 0.4605 0.5457 0.4605 0.6786
No log 8.0 440 0.4757 0.5501 0.4757 0.6897
No log 8.0364 442 0.5000 0.5324 0.5000 0.7071
No log 8.0727 444 0.4731 0.5632 0.4731 0.6878
No log 8.1091 446 0.4896 0.6182 0.4896 0.6997
No log 8.1455 448 0.6219 0.4686 0.6219 0.7886
No log 8.1818 450 0.6925 0.5103 0.6925 0.8321
No log 8.2182 452 0.6557 0.5455 0.6557 0.8097
No log 8.2545 454 0.5561 0.5358 0.5561 0.7457
No log 8.2909 456 0.5323 0.5822 0.5323 0.7296
No log 8.3273 458 0.5136 0.6317 0.5136 0.7167
No log 8.3636 460 0.4810 0.6184 0.4810 0.6935
No log 8.4 462 0.4895 0.5797 0.4895 0.6996
No log 8.4364 464 0.4971 0.5707 0.4971 0.7051
No log 8.4727 466 0.5096 0.5707 0.5096 0.7139
No log 8.5091 468 0.5379 0.5999 0.5379 0.7334
No log 8.5455 470 0.5578 0.5869 0.5578 0.7469
No log 8.5818 472 0.5314 0.5414 0.5314 0.7290
No log 8.6182 474 0.4997 0.5227 0.4997 0.7069
No log 8.6545 476 0.5035 0.5768 0.5035 0.7096
No log 8.6909 478 0.5029 0.5897 0.5029 0.7092
No log 8.7273 480 0.5025 0.5672 0.5025 0.7089
No log 8.7636 482 0.5720 0.5779 0.5720 0.7563
No log 8.8 484 0.6173 0.5061 0.6173 0.7857
No log 8.8364 486 0.5904 0.5358 0.5904 0.7684
No log 8.8727 488 0.5865 0.5538 0.5865 0.7658
No log 8.9091 490 0.5721 0.5387 0.5721 0.7564
No log 8.9455 492 0.5588 0.5929 0.5588 0.7475
No log 8.9818 494 0.5832 0.4948 0.5832 0.7637
No log 9.0182 496 0.5877 0.5286 0.5877 0.7666
No log 9.0545 498 0.5768 0.5174 0.5768 0.7594
0.3323 9.0909 500 0.5834 0.5594 0.5834 0.7638
0.3323 9.1273 502 0.5793 0.6210 0.5793 0.7611
0.3323 9.1636 504 0.5758 0.6210 0.5758 0.7588
0.3323 9.2 506 0.6177 0.4573 0.6177 0.7860
0.3323 9.2364 508 0.6310 0.4525 0.6310 0.7944
0.3323 9.2727 510 0.6204 0.4210 0.6204 0.7876
0.3323 9.3091 512 0.6400 0.4140 0.6400 0.8000
0.3323 9.3455 514 0.6372 0.4282 0.6372 0.7982

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k16_task7_organization

Finetuned
(4019)
this model