ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k14_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5523
  • Qwk: 0.4585
  • Mse: 0.5523
  • Rmse: 0.7432

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0417 2 2.4626 -0.0646 2.4626 1.5693
No log 0.0833 4 1.2253 0.1274 1.2253 1.1069
No log 0.125 6 0.8162 0.0949 0.8162 0.9034
No log 0.1667 8 0.7455 0.1321 0.7455 0.8634
No log 0.2083 10 0.7423 0.1770 0.7423 0.8616
No log 0.25 12 0.7920 0.1724 0.7920 0.8899
No log 0.2917 14 0.7724 0.1724 0.7724 0.8789
No log 0.3333 16 0.7707 0.0944 0.7707 0.8779
No log 0.375 18 0.7666 0.0944 0.7666 0.8756
No log 0.4167 20 0.7308 0.0851 0.7308 0.8549
No log 0.4583 22 0.7902 0.1184 0.7902 0.8889
No log 0.5 24 0.6786 0.1093 0.6786 0.8238
No log 0.5417 26 0.6495 0.1660 0.6495 0.8059
No log 0.5833 28 0.8008 0.2046 0.8008 0.8949
No log 0.625 30 0.7533 0.1711 0.7533 0.8679
No log 0.6667 32 0.7311 0.0947 0.7311 0.8550
No log 0.7083 34 0.6677 0.1327 0.6677 0.8172
No log 0.75 36 0.7408 0.0940 0.7408 0.8607
No log 0.7917 38 0.8065 0.1352 0.8065 0.8980
No log 0.8333 40 0.9081 0.2408 0.9081 0.9529
No log 0.875 42 0.8494 0.2109 0.8494 0.9216
No log 0.9167 44 0.7422 0.1365 0.7422 0.8615
No log 0.9583 46 0.7182 0.1352 0.7182 0.8474
No log 1.0 48 0.6215 0.1646 0.6215 0.7883
No log 1.0417 50 0.5927 0.3980 0.5927 0.7699
No log 1.0833 52 0.5870 0.3980 0.5870 0.7662
No log 1.125 54 0.5832 0.4432 0.5832 0.7637
No log 1.1667 56 0.6393 0.3693 0.6393 0.7996
No log 1.2083 58 0.6606 0.3371 0.6606 0.8127
No log 1.25 60 0.6267 0.2776 0.6267 0.7917
No log 1.2917 62 0.5668 0.3690 0.5668 0.7529
No log 1.3333 64 0.5560 0.4206 0.5560 0.7456
No log 1.375 66 0.5510 0.3915 0.5510 0.7423
No log 1.4167 68 0.5596 0.4338 0.5596 0.7481
No log 1.4583 70 0.5916 0.4776 0.5916 0.7692
No log 1.5 72 0.7387 0.4466 0.7387 0.8595
No log 1.5417 74 0.6824 0.4631 0.6824 0.8261
No log 1.5833 76 0.5484 0.4898 0.5484 0.7405
No log 1.625 78 0.5943 0.4243 0.5943 0.7709
No log 1.6667 80 0.6476 0.3996 0.6476 0.8047
No log 1.7083 82 0.6146 0.3494 0.6146 0.7840
No log 1.75 84 0.5809 0.4161 0.5809 0.7622
No log 1.7917 86 0.6449 0.4528 0.6449 0.8030
No log 1.8333 88 0.7007 0.4030 0.7007 0.8371
No log 1.875 90 0.6790 0.3783 0.6790 0.8240
No log 1.9167 92 0.6047 0.3489 0.6047 0.7776
No log 1.9583 94 0.6327 0.3267 0.6327 0.7955
No log 2.0 96 0.6377 0.3267 0.6377 0.7985
No log 2.0417 98 0.6131 0.3781 0.6131 0.7830
No log 2.0833 100 0.6226 0.4276 0.6226 0.7891
No log 2.125 102 0.6305 0.4441 0.6305 0.7940
No log 2.1667 104 0.6270 0.5057 0.6270 0.7918
No log 2.2083 106 0.6663 0.4373 0.6663 0.8163
No log 2.25 108 0.7054 0.3572 0.7054 0.8399
No log 2.2917 110 0.6452 0.4425 0.6452 0.8033
No log 2.3333 112 0.6270 0.4743 0.6270 0.7918
No log 2.375 114 0.6372 0.4980 0.6372 0.7983
No log 2.4167 116 0.6384 0.5710 0.6384 0.7990
No log 2.4583 118 0.6419 0.5379 0.6419 0.8012
No log 2.5 120 0.6258 0.5663 0.6258 0.7911
No log 2.5417 122 0.6164 0.5979 0.6164 0.7851
No log 2.5833 124 0.6072 0.5389 0.6072 0.7793
No log 2.625 126 0.6083 0.5235 0.6083 0.7800
No log 2.6667 128 0.5955 0.5167 0.5955 0.7717
No log 2.7083 130 0.5907 0.5167 0.5907 0.7685
No log 2.75 132 0.5849 0.5235 0.5849 0.7648
No log 2.7917 134 0.6066 0.5220 0.6066 0.7788
No log 2.8333 136 1.0885 0.2725 1.0885 1.0433
No log 2.875 138 1.2737 0.2421 1.2737 1.1286
No log 2.9167 140 0.9711 0.2023 0.9711 0.9855
No log 2.9583 142 0.6557 0.4575 0.6557 0.8098
No log 3.0 144 0.5581 0.4613 0.5581 0.7470
No log 3.0417 146 0.5557 0.4726 0.5557 0.7455
No log 3.0833 148 0.5757 0.4979 0.5757 0.7587
No log 3.125 150 0.8967 0.2981 0.8967 0.9469
No log 3.1667 152 1.1043 0.2263 1.1043 1.0508
No log 3.2083 154 0.9055 0.2772 0.9055 0.9516
No log 3.25 156 0.6130 0.4507 0.6130 0.7829
No log 3.2917 158 0.5787 0.4378 0.5787 0.7607
No log 3.3333 160 0.5945 0.4795 0.5945 0.7710
No log 3.375 162 0.5553 0.4878 0.5553 0.7452
No log 3.4167 164 0.6376 0.4868 0.6376 0.7985
No log 3.4583 166 0.6703 0.4732 0.6703 0.8187
No log 3.5 168 0.5831 0.4656 0.5831 0.7636
No log 3.5417 170 0.6275 0.4875 0.6275 0.7922
No log 3.5833 172 0.7071 0.4371 0.7071 0.8409
No log 3.625 174 0.6593 0.4795 0.6593 0.8120
No log 3.6667 176 0.6096 0.4795 0.6096 0.7808
No log 3.7083 178 0.5789 0.4726 0.5789 0.7608
No log 3.75 180 0.6086 0.4850 0.6086 0.7801
No log 3.7917 182 0.6472 0.5030 0.6472 0.8045
No log 3.8333 184 0.5750 0.5494 0.5750 0.7583
No log 3.875 186 0.5751 0.5437 0.5751 0.7583
No log 3.9167 188 0.7346 0.4367 0.7346 0.8571
No log 3.9583 190 0.6605 0.4755 0.6605 0.8127
No log 4.0 192 0.5744 0.5087 0.5744 0.7579
No log 4.0417 194 0.7858 0.4558 0.7858 0.8864
No log 4.0833 196 0.8056 0.4260 0.8056 0.8975
No log 4.125 198 0.6075 0.5233 0.6075 0.7794
No log 4.1667 200 0.5348 0.5476 0.5348 0.7313
No log 4.2083 202 0.5258 0.5656 0.5258 0.7252
No log 4.25 204 0.6271 0.5073 0.6271 0.7919
No log 4.2917 206 0.7010 0.4584 0.7010 0.8373
No log 4.3333 208 0.6096 0.5131 0.6096 0.7807
No log 4.375 210 0.5180 0.6184 0.5180 0.7197
No log 4.4167 212 0.5138 0.6184 0.5138 0.7168
No log 4.4583 214 0.5543 0.6259 0.5543 0.7445
No log 4.5 216 0.6736 0.4836 0.6736 0.8207
No log 4.5417 218 0.6220 0.4930 0.6220 0.7886
No log 4.5833 220 0.5197 0.6118 0.5197 0.7209
No log 4.625 222 0.5116 0.5604 0.5116 0.7152
No log 4.6667 224 0.5332 0.6269 0.5332 0.7302
No log 4.7083 226 0.5504 0.6259 0.5504 0.7419
No log 4.75 228 0.5584 0.5991 0.5584 0.7473
No log 4.7917 230 0.5351 0.5205 0.5351 0.7315
No log 4.8333 232 0.5322 0.4937 0.5322 0.7295
No log 4.875 234 0.5325 0.5167 0.5325 0.7297
No log 4.9167 236 0.5457 0.5205 0.5457 0.7387
No log 4.9583 238 0.6550 0.4890 0.6550 0.8093
No log 5.0 240 0.6743 0.4836 0.6743 0.8212
No log 5.0417 242 0.5783 0.5767 0.5783 0.7605
No log 5.0833 244 0.5483 0.5516 0.5483 0.7405
No log 5.125 246 0.5921 0.4925 0.5921 0.7695
No log 5.1667 248 0.5676 0.5271 0.5676 0.7534
No log 5.2083 250 0.5364 0.5379 0.5364 0.7324
No log 5.25 252 0.5295 0.5379 0.5295 0.7277
No log 5.2917 254 0.5325 0.5739 0.5325 0.7298
No log 5.3333 256 0.5503 0.5479 0.5503 0.7418
No log 5.375 258 0.5677 0.5671 0.5677 0.7534
No log 5.4167 260 0.6274 0.4997 0.6274 0.7921
No log 5.4583 262 0.5828 0.5271 0.5828 0.7634
No log 5.5 264 0.5464 0.4831 0.5464 0.7392
No log 5.5417 266 0.5607 0.5190 0.5607 0.7488
No log 5.5833 268 0.5385 0.4795 0.5385 0.7338
No log 5.625 270 0.5492 0.5621 0.5492 0.7411
No log 5.6667 272 0.5331 0.5841 0.5331 0.7301
No log 5.7083 274 0.5518 0.5621 0.5518 0.7428
No log 5.75 276 0.5922 0.5636 0.5922 0.7695
No log 5.7917 278 0.6077 0.5636 0.6077 0.7795
No log 5.8333 280 0.5429 0.5918 0.5429 0.7368
No log 5.875 282 0.5397 0.4984 0.5397 0.7346
No log 5.9167 284 0.6051 0.5271 0.6051 0.7779
No log 5.9583 286 0.5928 0.5271 0.5928 0.7699
No log 6.0 288 0.5442 0.5271 0.5442 0.7377
No log 6.0417 290 0.5122 0.5413 0.5122 0.7157
No log 6.0833 292 0.6068 0.5061 0.6068 0.7790
No log 6.125 294 0.6065 0.4868 0.6065 0.7788
No log 6.1667 296 0.5089 0.5765 0.5089 0.7133
No log 6.2083 298 0.5056 0.5698 0.5056 0.7110
No log 6.25 300 0.5136 0.5698 0.5136 0.7166
No log 6.2917 302 0.4919 0.6001 0.4919 0.7013
No log 6.3333 304 0.5645 0.4825 0.5645 0.7514
No log 6.375 306 0.6009 0.4868 0.6009 0.7752
No log 6.4167 308 0.5777 0.5061 0.5777 0.7601
No log 6.4583 310 0.5439 0.5524 0.5439 0.7375
No log 6.5 312 0.6166 0.4821 0.6166 0.7852
No log 6.5417 314 0.6502 0.4635 0.6502 0.8064
No log 6.5833 316 0.5831 0.5122 0.5831 0.7636
No log 6.625 318 0.5420 0.4849 0.5420 0.7362
No log 6.6667 320 0.5692 0.4632 0.5692 0.7545
No log 6.7083 322 0.5989 0.4507 0.5989 0.7739
No log 6.75 324 0.5727 0.4821 0.5727 0.7568
No log 6.7917 326 0.5324 0.5405 0.5324 0.7296
No log 6.8333 328 0.5162 0.5386 0.5162 0.7185
No log 6.875 330 0.5072 0.6335 0.5072 0.7122
No log 6.9167 332 0.4950 0.6344 0.4950 0.7036
No log 6.9583 334 0.5003 0.5897 0.5003 0.7073
No log 7.0 336 0.5051 0.5899 0.5051 0.7107
No log 7.0417 338 0.5172 0.5756 0.5172 0.7192
No log 7.0833 340 0.5222 0.5683 0.5222 0.7226
No log 7.125 342 0.5143 0.6060 0.5143 0.7171
No log 7.1667 344 0.5184 0.4343 0.5184 0.7200
No log 7.2083 346 0.5272 0.4484 0.5272 0.7261
No log 7.25 348 0.5763 0.4997 0.5763 0.7592
No log 7.2917 350 0.5920 0.4594 0.5920 0.7694
No log 7.3333 352 0.5501 0.3906 0.5501 0.7417
No log 7.375 354 0.5375 0.5214 0.5375 0.7332
No log 7.4167 356 0.5897 0.5017 0.5897 0.7679
No log 7.4583 358 0.5639 0.5271 0.5639 0.7509
No log 7.5 360 0.5027 0.5687 0.5027 0.7090
No log 7.5417 362 0.5545 0.4868 0.5545 0.7447
No log 7.5833 364 0.5732 0.5095 0.5732 0.7571
No log 7.625 366 0.5122 0.5195 0.5122 0.7157
No log 7.6667 368 0.5011 0.5516 0.5011 0.7079
No log 7.7083 370 0.5383 0.5271 0.5383 0.7337
No log 7.75 372 0.5292 0.5271 0.5292 0.7275
No log 7.7917 374 0.5148 0.5348 0.5148 0.7175
No log 7.8333 376 0.5379 0.4631 0.5379 0.7334
No log 7.875 378 0.5316 0.4314 0.5316 0.7291
No log 7.9167 380 0.5219 0.4637 0.5219 0.7224
No log 7.9583 382 0.5482 0.5639 0.5482 0.7404
No log 8.0 384 0.6091 0.5677 0.6091 0.7804
No log 8.0417 386 0.5895 0.5774 0.5895 0.7678
No log 8.0833 388 0.5567 0.6063 0.5567 0.7461
No log 8.125 390 0.5259 0.6034 0.5259 0.7252
No log 8.1667 392 0.5125 0.5853 0.5125 0.7159
No log 8.2083 394 0.5005 0.6017 0.5005 0.7075
No log 8.25 396 0.5062 0.4891 0.5062 0.7115
No log 8.2917 398 0.5092 0.5267 0.5092 0.7136
No log 8.3333 400 0.5413 0.4788 0.5413 0.7357
No log 8.375 402 0.5542 0.5677 0.5542 0.7445
No log 8.4167 404 0.5111 0.5894 0.5111 0.7149
No log 8.4583 406 0.4714 0.6147 0.4714 0.6866
No log 8.5 408 0.4770 0.6584 0.4770 0.6906
No log 8.5417 410 0.5538 0.5061 0.5538 0.7442
No log 8.5833 412 0.5932 0.5266 0.5932 0.7702
No log 8.625 414 0.5872 0.5061 0.5872 0.7663
No log 8.6667 416 0.5265 0.5184 0.5265 0.7256
No log 8.7083 418 0.5110 0.5599 0.5110 0.7149
No log 8.75 420 0.4622 0.5343 0.4622 0.6798
No log 8.7917 422 0.4533 0.4538 0.4533 0.6733
No log 8.8333 424 0.4833 0.4912 0.4833 0.6952
No log 8.875 426 0.5475 0.4165 0.5475 0.7399
No log 8.9167 428 0.5752 0.4011 0.5752 0.7584
No log 8.9583 430 0.5419 0.4518 0.5419 0.7362
No log 9.0 432 0.4985 0.4881 0.4985 0.7060
No log 9.0417 434 0.4846 0.4314 0.4846 0.6961
No log 9.0833 436 0.4753 0.4847 0.4753 0.6894
No log 9.125 438 0.4635 0.5714 0.4635 0.6808
No log 9.1667 440 0.4876 0.5495 0.4876 0.6983
No log 9.2083 442 0.5325 0.5014 0.5325 0.7297
No log 9.25 444 0.5330 0.5414 0.5330 0.7301
No log 9.2917 446 0.5255 0.5483 0.5255 0.7249
No log 9.3333 448 0.5339 0.5692 0.5339 0.7307
No log 9.375 450 0.5288 0.5267 0.5288 0.7272
No log 9.4167 452 0.5383 0.5061 0.5383 0.7337
No log 9.4583 454 0.5257 0.5061 0.5257 0.7250
No log 9.5 456 0.5158 0.5127 0.5158 0.7182
No log 9.5417 458 0.4922 0.5143 0.4922 0.7016
No log 9.5833 460 0.5118 0.5271 0.5118 0.7154
No log 9.625 462 0.5482 0.5677 0.5482 0.7404
No log 9.6667 464 0.5250 0.5817 0.5250 0.7246
No log 9.7083 466 0.5035 0.4973 0.5035 0.7096
No log 9.75 468 0.5433 0.5483 0.5433 0.7371
No log 9.7917 470 0.5617 0.5283 0.5617 0.7495
No log 9.8333 472 0.5516 0.5571 0.5516 0.7427
No log 9.875 474 0.5226 0.5098 0.5226 0.7229
No log 9.9167 476 0.4919 0.5203 0.4919 0.7014
No log 9.9583 478 0.4910 0.5386 0.4910 0.7007
No log 10.0 480 0.4934 0.5141 0.4934 0.7024
No log 10.0417 482 0.5022 0.5943 0.5022 0.7086
No log 10.0833 484 0.5367 0.5352 0.5367 0.7326
No log 10.125 486 0.5743 0.5161 0.5743 0.7578
No log 10.1667 488 0.5535 0.4931 0.5535 0.7439
No log 10.2083 490 0.5061 0.5781 0.5061 0.7114
No log 10.25 492 0.4943 0.5970 0.4943 0.7031
No log 10.2917 494 0.5020 0.5649 0.5020 0.7085
No log 10.3333 496 0.4956 0.5649 0.4956 0.7040
No log 10.375 498 0.4934 0.5649 0.4934 0.7024
0.2898 10.4167 500 0.5483 0.5078 0.5483 0.7405
0.2898 10.4583 502 0.5907 0.4886 0.5907 0.7686
0.2898 10.5 504 0.5425 0.4913 0.5425 0.7365
0.2898 10.5417 506 0.5311 0.4677 0.5311 0.7287
0.2898 10.5833 508 0.5375 0.4314 0.5375 0.7331
0.2898 10.625 510 0.5523 0.4585 0.5523 0.7432

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k14_task7_organization

Finetuned
(4019)
this model