ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5056
  • Qwk: 0.4726
  • Mse: 0.5056
  • Rmse: 0.7111

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0645 2 2.5893 -0.0593 2.5893 1.6091
No log 0.1290 4 1.2040 0.2129 1.2040 1.0973
No log 0.1935 6 0.9256 -0.0672 0.9256 0.9621
No log 0.2581 8 1.1536 -0.0234 1.1536 1.0741
No log 0.3226 10 1.2013 0.0728 1.2013 1.0961
No log 0.3871 12 0.9399 0.2142 0.9399 0.9695
No log 0.4516 14 0.8336 0.1972 0.8336 0.9130
No log 0.5161 16 0.7659 0.0798 0.7659 0.8751
No log 0.5806 18 0.7496 0.1139 0.7496 0.8658
No log 0.6452 20 0.7991 0.3302 0.7991 0.8939
No log 0.7097 22 0.7774 0.2227 0.7774 0.8817
No log 0.7742 24 0.7498 0.0898 0.7498 0.8659
No log 0.8387 26 0.8412 0.2066 0.8412 0.9172
No log 0.9032 28 0.9090 0.2942 0.9090 0.9534
No log 0.9677 30 1.1534 0.1786 1.1534 1.0740
No log 1.0323 32 1.3030 0.0299 1.3030 1.1415
No log 1.0968 34 1.1791 0.1246 1.1791 1.0859
No log 1.1613 36 1.0859 0.2081 1.0859 1.0421
No log 1.2258 38 0.9872 0.3085 0.9872 0.9936
No log 1.2903 40 0.8256 0.2613 0.8256 0.9086
No log 1.3548 42 0.7827 0.1626 0.7827 0.8847
No log 1.4194 44 0.9088 0.3466 0.9088 0.9533
No log 1.4839 46 0.8579 0.2942 0.8579 0.9262
No log 1.5484 48 0.7021 0.2355 0.7021 0.8379
No log 1.6129 50 0.6411 0.3482 0.6411 0.8007
No log 1.6774 52 0.7054 0.4154 0.7054 0.8399
No log 1.7419 54 0.6641 0.4369 0.6641 0.8150
No log 1.8065 56 0.6371 0.4486 0.6371 0.7982
No log 1.8710 58 0.6951 0.3948 0.6951 0.8338
No log 1.9355 60 0.6815 0.3807 0.6815 0.8255
No log 2.0 62 0.7929 0.3640 0.7929 0.8904
No log 2.0645 64 0.8079 0.3873 0.8079 0.8988
No log 2.1290 66 0.7280 0.4016 0.7280 0.8532
No log 2.1935 68 0.7568 0.3523 0.7568 0.8699
No log 2.2581 70 0.7191 0.3417 0.7191 0.8480
No log 2.3226 72 0.7497 0.3183 0.7497 0.8659
No log 2.3871 74 0.8145 0.2832 0.8145 0.9025
No log 2.4516 76 0.6962 0.3011 0.6962 0.8344
No log 2.5161 78 0.6695 0.2509 0.6695 0.8182
No log 2.5806 80 0.6421 0.3100 0.6421 0.8013
No log 2.6452 82 0.7290 0.4351 0.7290 0.8538
No log 2.7097 84 0.7579 0.3434 0.7579 0.8706
No log 2.7742 86 0.6403 0.4589 0.6403 0.8002
No log 2.8387 88 0.6176 0.4127 0.6176 0.7858
No log 2.9032 90 0.6364 0.4179 0.6364 0.7978
No log 2.9677 92 0.6437 0.3937 0.6437 0.8023
No log 3.0323 94 0.5828 0.3910 0.5828 0.7634
No log 3.0968 96 0.6253 0.5422 0.6253 0.7908
No log 3.1613 98 0.5866 0.4182 0.5866 0.7659
No log 3.2258 100 0.6149 0.4257 0.6149 0.7841
No log 3.2903 102 0.6251 0.4157 0.6251 0.7907
No log 3.3548 104 0.5839 0.4816 0.5839 0.7641
No log 3.4194 106 0.5811 0.3780 0.5811 0.7623
No log 3.4839 108 0.5862 0.3781 0.5862 0.7656
No log 3.5484 110 0.7447 0.3643 0.7447 0.8630
No log 3.6129 112 0.8448 0.3287 0.8448 0.9192
No log 3.6774 114 0.7281 0.3918 0.7281 0.8533
No log 3.7419 116 0.6343 0.3041 0.6343 0.7964
No log 3.8065 118 0.6757 0.3729 0.6757 0.8220
No log 3.8710 120 0.7076 0.4250 0.7076 0.8412
No log 3.9355 122 0.6450 0.4128 0.6450 0.8031
No log 4.0 124 0.6412 0.4402 0.6412 0.8008
No log 4.0645 126 0.6318 0.4715 0.6318 0.7949
No log 4.1290 128 0.7154 0.3892 0.7154 0.8458
No log 4.1935 130 0.7088 0.4003 0.7088 0.8419
No log 4.2581 132 0.6216 0.5127 0.6216 0.7884
No log 4.3226 134 0.5800 0.4547 0.5800 0.7616
No log 4.3871 136 0.5818 0.4726 0.5818 0.7628
No log 4.4516 138 0.5658 0.4591 0.5658 0.7522
No log 4.5161 140 0.6075 0.4770 0.6075 0.7794
No log 4.5806 142 0.6305 0.4997 0.6305 0.7941
No log 4.6452 144 0.5699 0.5734 0.5699 0.7549
No log 4.7097 146 0.6026 0.4847 0.6026 0.7763
No log 4.7742 148 0.5924 0.5430 0.5924 0.7697
No log 4.8387 150 0.5934 0.4801 0.5934 0.7704
No log 4.9032 152 0.6669 0.4625 0.6669 0.8166
No log 4.9677 154 0.9280 0.3418 0.9280 0.9633
No log 5.0323 156 0.9887 0.3317 0.9887 0.9943
No log 5.0968 158 0.7490 0.3786 0.7490 0.8655
No log 5.1613 160 0.5639 0.4985 0.5639 0.7510
No log 5.2258 162 0.5687 0.5627 0.5687 0.7541
No log 5.2903 164 0.5673 0.5348 0.5673 0.7532
No log 5.3548 166 0.5381 0.5692 0.5381 0.7336
No log 5.4194 168 0.5446 0.5640 0.5446 0.7380
No log 5.4839 170 0.5364 0.5739 0.5364 0.7324
No log 5.5484 172 0.5156 0.5037 0.5156 0.7180
No log 5.6129 174 0.5032 0.5363 0.5032 0.7093
No log 5.6774 176 0.4897 0.4910 0.4897 0.6998
No log 5.7419 178 0.5031 0.6182 0.5031 0.7093
No log 5.8065 180 0.5316 0.6030 0.5316 0.7291
No log 5.8710 182 0.5921 0.6042 0.5921 0.7695
No log 5.9355 184 0.7775 0.5160 0.7775 0.8818
No log 6.0 186 0.8226 0.4074 0.8226 0.9070
No log 6.0645 188 0.6494 0.6156 0.6494 0.8058
No log 6.1290 190 0.6492 0.6156 0.6492 0.8057
No log 6.1935 192 0.7373 0.5160 0.7373 0.8587
No log 6.2581 194 0.6655 0.5657 0.6655 0.8158
No log 6.3226 196 0.5303 0.5553 0.5303 0.7282
No log 6.3871 198 0.4911 0.5476 0.4911 0.7008
No log 6.4516 200 0.4943 0.5053 0.4943 0.7031
No log 6.5161 202 0.5084 0.5015 0.5084 0.7130
No log 6.5806 204 0.5022 0.4838 0.5022 0.7086
No log 6.6452 206 0.4980 0.5711 0.4980 0.7057
No log 6.7097 208 0.5302 0.4929 0.5302 0.7281
No log 6.7742 210 0.5218 0.4929 0.5218 0.7223
No log 6.8387 212 0.5060 0.5289 0.5060 0.7113
No log 6.9032 214 0.5121 0.5492 0.5121 0.7156
No log 6.9677 216 0.5169 0.5565 0.5169 0.7189
No log 7.0323 218 0.5543 0.4595 0.5543 0.7445
No log 7.0968 220 0.5501 0.4595 0.5501 0.7417
No log 7.1613 222 0.5232 0.5053 0.5232 0.7233
No log 7.2258 224 0.6161 0.4749 0.6161 0.7849
No log 7.2903 226 0.5910 0.4769 0.5910 0.7687
No log 7.3548 228 0.5217 0.5797 0.5217 0.7223
No log 7.4194 230 0.5411 0.4813 0.5411 0.7356
No log 7.4839 232 0.5182 0.4615 0.5182 0.7199
No log 7.5484 234 0.5179 0.4809 0.5179 0.7197
No log 7.6129 236 0.5925 0.4576 0.5925 0.7697
No log 7.6774 238 0.6424 0.4409 0.6424 0.8015
No log 7.7419 240 0.5669 0.4774 0.5669 0.7529
No log 7.8065 242 0.5555 0.4149 0.5555 0.7453
No log 7.8710 244 0.6216 0.4715 0.6216 0.7884
No log 7.9355 246 0.5509 0.4942 0.5509 0.7422
No log 8.0 248 0.5421 0.4816 0.5421 0.7363
No log 8.0645 250 0.5658 0.4964 0.5658 0.7522
No log 8.1290 252 0.5376 0.5022 0.5376 0.7332
No log 8.1935 254 0.5462 0.5751 0.5462 0.7390
No log 8.2581 256 0.6190 0.5030 0.6190 0.7868
No log 8.3226 258 0.6005 0.4616 0.6005 0.7749
No log 8.3871 260 0.6228 0.4778 0.6228 0.7892
No log 8.4516 262 0.5415 0.5809 0.5415 0.7359
No log 8.5161 264 0.5130 0.5813 0.5130 0.7162
No log 8.5806 266 0.5246 0.5223 0.5246 0.7243
No log 8.6452 268 0.5075 0.5159 0.5075 0.7124
No log 8.7097 270 0.5092 0.4743 0.5092 0.7136
No log 8.7742 272 0.5637 0.5063 0.5637 0.7508
No log 8.8387 274 0.6236 0.52 0.6236 0.7897
No log 8.9032 276 0.5399 0.5576 0.5399 0.7348
No log 8.9677 278 0.5047 0.5151 0.5047 0.7104
No log 9.0323 280 0.5751 0.5460 0.5751 0.7583
No log 9.0968 282 0.5863 0.5460 0.5863 0.7657
No log 9.1613 284 0.5109 0.5136 0.5109 0.7148
No log 9.2258 286 0.5046 0.5753 0.5046 0.7104
No log 9.2903 288 0.5168 0.5692 0.5168 0.7189
No log 9.3548 290 0.5152 0.5697 0.5152 0.7177
No log 9.4194 292 0.6059 0.5061 0.6059 0.7784
No log 9.4839 294 0.7343 0.4844 0.7343 0.8569
No log 9.5484 296 0.7735 0.4496 0.7735 0.8795
No log 9.6129 298 0.6492 0.4811 0.6492 0.8057
No log 9.6774 300 0.5312 0.5621 0.5312 0.7288
No log 9.7419 302 0.5246 0.5621 0.5246 0.7243
No log 9.8065 304 0.5361 0.5498 0.5361 0.7322
No log 9.8710 306 0.6014 0.5543 0.6014 0.7755
No log 9.9355 308 0.6081 0.4967 0.6081 0.7798
No log 10.0 310 0.5427 0.4150 0.5427 0.7367
No log 10.0645 312 0.5850 0.4569 0.5850 0.7649
No log 10.1290 314 0.6970 0.4144 0.6970 0.8349
No log 10.1935 316 0.7785 0.3777 0.7785 0.8823
No log 10.2581 318 0.7112 0.4144 0.7112 0.8433
No log 10.3226 320 0.5961 0.3868 0.5961 0.7720
No log 10.3871 322 0.5652 0.3336 0.5652 0.7518
No log 10.4516 324 0.5708 0.3993 0.5708 0.7555
No log 10.5161 326 0.5398 0.4746 0.5398 0.7347
No log 10.5806 328 0.5197 0.4634 0.5197 0.7209
No log 10.6452 330 0.5416 0.4925 0.5416 0.7359
No log 10.7097 332 0.5503 0.4855 0.5503 0.7418
No log 10.7742 334 0.5135 0.4726 0.5135 0.7166
No log 10.8387 336 0.5163 0.4569 0.5163 0.7185
No log 10.9032 338 0.5165 0.3728 0.5165 0.7187
No log 10.9677 340 0.5228 0.4918 0.5228 0.7230
No log 11.0323 342 0.5113 0.4299 0.5113 0.7151
No log 11.0968 344 0.4980 0.5214 0.4980 0.7057
No log 11.1613 346 0.4922 0.5455 0.4922 0.7016
No log 11.2258 348 0.4884 0.5567 0.4884 0.6988
No log 11.2903 350 0.4889 0.5232 0.4889 0.6992
No log 11.3548 352 0.4894 0.5307 0.4894 0.6996
No log 11.4194 354 0.4964 0.4849 0.4964 0.7046
No log 11.4839 356 0.5097 0.5266 0.5097 0.7139
No log 11.5484 358 0.5179 0.4913 0.5179 0.7196
No log 11.6129 360 0.5062 0.5352 0.5062 0.7115
No log 11.6774 362 0.4840 0.5373 0.4840 0.6957
No log 11.7419 364 0.4892 0.5611 0.4892 0.6994
No log 11.8065 366 0.5281 0.5498 0.5281 0.7267
No log 11.8710 368 0.6122 0.5347 0.6122 0.7824
No log 11.9355 370 0.7462 0.4684 0.7462 0.8638
No log 12.0 372 0.6962 0.4844 0.6962 0.8344
No log 12.0645 374 0.5597 0.5400 0.5597 0.7481
No log 12.1290 376 0.4926 0.5678 0.4926 0.7019
No log 12.1935 378 0.4880 0.5092 0.4880 0.6985
No log 12.2581 380 0.4919 0.4575 0.4919 0.7014
No log 12.3226 382 0.4774 0.5340 0.4774 0.6909
No log 12.3871 384 0.4799 0.5340 0.4799 0.6927
No log 12.4516 386 0.4839 0.4547 0.4839 0.6956
No log 12.5161 388 0.4955 0.4656 0.4955 0.7039
No log 12.5806 390 0.4947 0.4136 0.4947 0.7034
No log 12.6452 392 0.4955 0.4052 0.4955 0.7039
No log 12.7097 394 0.5027 0.4291 0.5027 0.7090
No log 12.7742 396 0.5083 0.4291 0.5083 0.7129
No log 12.8387 398 0.5122 0.4291 0.5122 0.7157
No log 12.9032 400 0.5060 0.3995 0.5060 0.7113
No log 12.9677 402 0.5137 0.5010 0.5137 0.7167
No log 13.0323 404 0.5226 0.5666 0.5226 0.7229
No log 13.0968 406 0.4961 0.4762 0.4961 0.7043
No log 13.1613 408 0.5275 0.5357 0.5275 0.7263
No log 13.2258 410 0.5786 0.4887 0.5786 0.7607
No log 13.2903 412 0.5455 0.4663 0.5455 0.7385
No log 13.3548 414 0.5032 0.4378 0.5032 0.7094
No log 13.4194 416 0.5129 0.3728 0.5129 0.7162
No log 13.4839 418 0.5124 0.4111 0.5124 0.7158
No log 13.5484 420 0.5090 0.4962 0.5090 0.7134
No log 13.6129 422 0.5250 0.4867 0.5250 0.7246
No log 13.6774 424 0.5577 0.5017 0.5577 0.7468
No log 13.7419 426 0.5377 0.4774 0.5377 0.7333
No log 13.8065 428 0.4988 0.4888 0.4988 0.7063
No log 13.8710 430 0.5055 0.4101 0.5055 0.7110
No log 13.9355 432 0.4975 0.4656 0.4975 0.7053
No log 14.0 434 0.4858 0.4504 0.4858 0.6970
No log 14.0645 436 0.4946 0.4888 0.4946 0.7033
No log 14.1290 438 0.4874 0.4888 0.4874 0.6982
No log 14.1935 440 0.4860 0.4526 0.4860 0.6972
No log 14.2581 442 0.5063 0.5098 0.5063 0.7115
No log 14.3226 444 0.5073 0.5098 0.5073 0.7122
No log 14.3871 446 0.4765 0.4788 0.4765 0.6903
No log 14.4516 448 0.4740 0.5404 0.4740 0.6885
No log 14.5161 450 0.4754 0.5533 0.4754 0.6895
No log 14.5806 452 0.4783 0.5361 0.4783 0.6916
No log 14.6452 454 0.4821 0.5286 0.4821 0.6944
No log 14.7097 456 0.4787 0.5361 0.4787 0.6919
No log 14.7742 458 0.4861 0.5289 0.4861 0.6972
No log 14.8387 460 0.5208 0.5086 0.5208 0.7217
No log 14.9032 462 0.5107 0.5086 0.5107 0.7147
No log 14.9677 464 0.4773 0.5703 0.4773 0.6908
No log 15.0323 466 0.4896 0.5678 0.4896 0.6997
No log 15.0968 468 0.4995 0.6248 0.4995 0.7067
No log 15.1613 470 0.4891 0.5881 0.4891 0.6993
No log 15.2258 472 0.4752 0.6541 0.4752 0.6894
No log 15.2903 474 0.4781 0.6418 0.4781 0.6915
No log 15.3548 476 0.4652 0.6426 0.4652 0.6820
No log 15.4194 478 0.4701 0.5373 0.4701 0.6856
No log 15.4839 480 0.4693 0.5665 0.4693 0.6851
No log 15.5484 482 0.4678 0.5883 0.4678 0.6840
No log 15.6129 484 0.4645 0.5995 0.4645 0.6816
No log 15.6774 486 0.4741 0.6612 0.4741 0.6885
No log 15.7419 488 0.5272 0.5528 0.5272 0.7261
No log 15.8065 490 0.6510 0.5153 0.6510 0.8069
No log 15.8710 492 0.6472 0.5153 0.6472 0.8045
No log 15.9355 494 0.5300 0.5721 0.5300 0.7280
No log 16.0 496 0.4850 0.6530 0.4850 0.6964
No log 16.0645 498 0.5157 0.5920 0.5157 0.7182
0.3085 16.1290 500 0.4929 0.6214 0.4929 0.7021
0.3085 16.1935 502 0.4852 0.5114 0.4852 0.6965
0.3085 16.2581 504 0.4981 0.5591 0.4981 0.7057
0.3085 16.3226 506 0.4997 0.5343 0.4997 0.7069
0.3085 16.3871 508 0.4945 0.5133 0.4945 0.7032
0.3085 16.4516 510 0.5056 0.4726 0.5056 0.7111

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task7_organization

Finetuned
(4019)
this model