ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k8_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6133
  • Qwk: 0.5263
  • Mse: 0.6133
  • Rmse: 0.7831

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 2.5479 -0.1089 2.5479 1.5962
No log 0.1379 4 1.2177 0.1599 1.2177 1.1035
No log 0.2069 6 0.7706 0.0944 0.7706 0.8778
No log 0.2759 8 0.7456 0.0679 0.7456 0.8635
No log 0.3448 10 0.7414 -0.0027 0.7414 0.8611
No log 0.4138 12 0.7279 0.0 0.7279 0.8532
No log 0.4828 14 0.7766 0.0937 0.7766 0.8812
No log 0.5517 16 0.7200 0.0481 0.7200 0.8485
No log 0.6207 18 0.6502 0.1187 0.6502 0.8064
No log 0.6897 20 0.6595 0.1786 0.6595 0.8121
No log 0.7586 22 0.9164 0.3119 0.9164 0.9573
No log 0.8276 24 1.2603 0.1670 1.2603 1.1226
No log 0.8966 26 1.1666 0.2701 1.1666 1.0801
No log 0.9655 28 0.8904 0.3119 0.8904 0.9436
No log 1.0345 30 0.8173 0.2942 0.8173 0.9041
No log 1.1034 32 0.6870 0.3131 0.6870 0.8289
No log 1.1724 34 0.7652 0.2585 0.7652 0.8748
No log 1.2414 36 1.1959 0.1913 1.1959 1.0936
No log 1.3103 38 1.2546 0.1670 1.2546 1.1201
No log 1.3793 40 0.7925 0.4162 0.7925 0.8902
No log 1.4483 42 0.5437 0.3866 0.5437 0.7373
No log 1.5172 44 0.6455 0.4731 0.6455 0.8034
No log 1.5862 46 0.6284 0.4479 0.6284 0.7927
No log 1.6552 48 0.5636 0.2024 0.5636 0.7507
No log 1.7241 50 0.7783 0.3090 0.7783 0.8822
No log 1.7931 52 0.8302 0.3354 0.8302 0.9112
No log 1.8621 54 0.7651 0.1365 0.7651 0.8747
No log 1.9310 56 0.6960 0.0937 0.6960 0.8342
No log 2.0 58 0.6095 0.3716 0.6095 0.7807
No log 2.0690 60 0.6056 0.3546 0.6056 0.7782
No log 2.1379 62 0.5992 0.3092 0.5992 0.7741
No log 2.2069 64 0.5591 0.4949 0.5591 0.7477
No log 2.2759 66 0.6310 0.4250 0.6310 0.7944
No log 2.3448 68 0.7403 0.3933 0.7403 0.8604
No log 2.4138 70 0.6773 0.4219 0.6773 0.8230
No log 2.4828 72 0.5555 0.5079 0.5555 0.7453
No log 2.5517 74 0.5442 0.5386 0.5442 0.7377
No log 2.6207 76 0.5434 0.5227 0.5434 0.7371
No log 2.6897 78 0.5396 0.5003 0.5396 0.7346
No log 2.7586 80 0.5245 0.5203 0.5245 0.7242
No log 2.8276 82 0.5764 0.4434 0.5764 0.7592
No log 2.8966 84 0.6371 0.4341 0.6371 0.7982
No log 2.9655 86 0.6614 0.3921 0.6614 0.8133
No log 3.0345 88 0.6570 0.4021 0.6570 0.8106
No log 3.1034 90 0.5383 0.2822 0.5383 0.7337
No log 3.1724 92 0.5438 0.4420 0.5438 0.7374
No log 3.2414 94 0.6068 0.3950 0.6068 0.7790
No log 3.3103 96 0.5542 0.4685 0.5542 0.7444
No log 3.3793 98 0.5369 0.5161 0.5369 0.7327
No log 3.4483 100 0.5210 0.4724 0.5210 0.7218
No log 3.5172 102 0.5512 0.5195 0.5512 0.7425
No log 3.5862 104 0.5445 0.5127 0.5445 0.7379
No log 3.6552 106 0.5355 0.5600 0.5355 0.7318
No log 3.7241 108 0.5603 0.5923 0.5603 0.7485
No log 3.7931 110 0.5399 0.5719 0.5399 0.7348
No log 3.8621 112 0.7137 0.4279 0.7137 0.8448
No log 3.9310 114 0.9017 0.3321 0.9017 0.9496
No log 4.0 116 0.9484 0.3309 0.9484 0.9739
No log 4.0690 118 0.7938 0.3499 0.7938 0.8910
No log 4.1379 120 0.6260 0.4985 0.6260 0.7912
No log 4.2069 122 0.5150 0.4448 0.5150 0.7176
No log 4.2759 124 0.5211 0.6154 0.5211 0.7219
No log 4.3448 126 0.5184 0.5897 0.5184 0.7200
No log 4.4138 128 0.5140 0.6305 0.5140 0.7170
No log 4.4828 130 0.5876 0.5373 0.5876 0.7666
No log 4.5517 132 0.6529 0.4703 0.6529 0.8080
No log 4.6207 134 0.6447 0.4756 0.6447 0.8029
No log 4.6897 136 0.5548 0.5748 0.5548 0.7448
No log 4.7586 138 0.4886 0.5123 0.4886 0.6990
No log 4.8276 140 0.4996 0.5420 0.4996 0.7068
No log 4.8966 142 0.4948 0.5867 0.4948 0.7034
No log 4.9655 144 0.4980 0.6143 0.4980 0.7057
No log 5.0345 146 0.6490 0.4511 0.6490 0.8056
No log 5.1034 148 0.6575 0.4462 0.6575 0.8109
No log 5.1724 150 0.5245 0.5445 0.5245 0.7242
No log 5.2414 152 0.4891 0.6506 0.4891 0.6994
No log 5.3103 154 0.5113 0.5237 0.5113 0.7151
No log 5.3793 156 0.5040 0.6452 0.5040 0.7100
No log 5.4483 158 0.5225 0.5414 0.5225 0.7229
No log 5.5172 160 0.5218 0.6034 0.5218 0.7224
No log 5.5862 162 0.5314 0.5233 0.5314 0.7290
No log 5.6552 164 0.5256 0.5621 0.5256 0.7250
No log 5.7241 166 0.5334 0.5621 0.5334 0.7304
No log 5.7931 168 0.5714 0.5721 0.5714 0.7559
No log 5.8621 170 0.6188 0.5073 0.6188 0.7866
No log 5.9310 172 0.7081 0.4717 0.7081 0.8415
No log 6.0 174 0.6237 0.5368 0.6237 0.7898
No log 6.0690 176 0.5860 0.5819 0.5860 0.7655
No log 6.1379 178 0.5856 0.6048 0.5856 0.7653
No log 6.2069 180 0.7151 0.4717 0.7151 0.8456
No log 6.2759 182 0.8312 0.3866 0.8312 0.9117
No log 6.3448 184 0.6986 0.4717 0.6986 0.8358
No log 6.4138 186 0.5791 0.6024 0.5791 0.7610
No log 6.4828 188 0.6039 0.5500 0.6039 0.7771
No log 6.5517 190 0.7818 0.3766 0.7818 0.8842
No log 6.6207 192 0.7817 0.3909 0.7817 0.8841
No log 6.6897 194 0.6728 0.4380 0.6728 0.8203
No log 6.7586 196 0.5891 0.5112 0.5891 0.7675
No log 6.8276 198 0.5303 0.5812 0.5303 0.7282
No log 6.8966 200 0.5408 0.5692 0.5408 0.7354
No log 6.9655 202 0.7172 0.4208 0.7172 0.8469
No log 7.0345 204 0.8857 0.3030 0.8857 0.9411
No log 7.1034 206 0.7271 0.4114 0.7271 0.8527
No log 7.1724 208 0.5514 0.5332 0.5514 0.7425
No log 7.2414 210 0.5604 0.5784 0.5604 0.7486
No log 7.3103 212 0.5853 0.5656 0.5853 0.7650
No log 7.3793 214 0.6293 0.5460 0.6293 0.7933
No log 7.4483 216 0.5841 0.5241 0.5841 0.7643
No log 7.5172 218 0.5788 0.4881 0.5788 0.7608
No log 7.5862 220 0.5585 0.4678 0.5585 0.7474
No log 7.6552 222 0.5528 0.5929 0.5528 0.7435
No log 7.7241 224 0.5645 0.5706 0.5645 0.7513
No log 7.7931 226 0.5976 0.4986 0.5976 0.7731
No log 7.8621 228 0.5486 0.5933 0.5486 0.7407
No log 7.9310 230 0.5135 0.6852 0.5135 0.7166
No log 8.0 232 0.5249 0.6503 0.5249 0.7245
No log 8.0690 234 0.6516 0.4794 0.6516 0.8072
No log 8.1379 236 0.7074 0.4427 0.7074 0.8411
No log 8.2069 238 0.6019 0.4794 0.6019 0.7759
No log 8.2759 240 0.5312 0.6616 0.5312 0.7288
No log 8.3448 242 0.5082 0.6210 0.5082 0.7129
No log 8.4138 244 0.5196 0.5900 0.5196 0.7209
No log 8.4828 246 0.5046 0.7041 0.5046 0.7104
No log 8.5517 248 0.6236 0.4794 0.6236 0.7897
No log 8.6207 250 0.6436 0.4794 0.6436 0.8023
No log 8.6897 252 0.5478 0.5315 0.5478 0.7401
No log 8.7586 254 0.4738 0.6769 0.4738 0.6883
No log 8.8276 256 0.4659 0.6673 0.4659 0.6825
No log 8.8966 258 0.4811 0.6317 0.4811 0.6936
No log 8.9655 260 0.4977 0.6367 0.4977 0.7055
No log 9.0345 262 0.5395 0.5552 0.5395 0.7345
No log 9.1034 264 0.5195 0.6367 0.5195 0.7208
No log 9.1724 266 0.5203 0.6367 0.5203 0.7213
No log 9.2414 268 0.4720 0.6616 0.4720 0.6870
No log 9.3103 270 0.4694 0.6870 0.4694 0.6851
No log 9.3793 272 0.4768 0.6694 0.4768 0.6905
No log 9.4483 274 0.5184 0.5599 0.5184 0.7200
No log 9.5172 276 0.5056 0.5208 0.5056 0.7110
No log 9.5862 278 0.4904 0.4902 0.4904 0.7003
No log 9.6552 280 0.5352 0.5189 0.5352 0.7316
No log 9.7241 282 0.5662 0.4835 0.5662 0.7525
No log 9.7931 284 0.5244 0.5291 0.5244 0.7241
No log 9.8621 286 0.4894 0.5902 0.4894 0.6996
No log 9.9310 288 0.5304 0.5636 0.5304 0.7283
No log 10.0 290 0.5536 0.5500 0.5536 0.7440
No log 10.0690 292 0.5671 0.5500 0.5671 0.7531
No log 10.1379 294 0.5232 0.5822 0.5232 0.7233
No log 10.2069 296 0.5079 0.5457 0.5079 0.7127
No log 10.2759 298 0.5095 0.5809 0.5095 0.7138
No log 10.3448 300 0.5315 0.6004 0.5315 0.7290
No log 10.4138 302 0.5023 0.6034 0.5023 0.7088
No log 10.4828 304 0.4965 0.6034 0.4965 0.7046
No log 10.5517 306 0.5042 0.6156 0.5042 0.7101
No log 10.6207 308 0.4841 0.6114 0.4841 0.6958
No log 10.6897 310 0.4893 0.6579 0.4893 0.6995
No log 10.7586 312 0.5151 0.6127 0.5151 0.7177
No log 10.8276 314 0.4906 0.6579 0.4906 0.7005
No log 10.8966 316 0.4825 0.5868 0.4825 0.6946
No log 10.9655 318 0.4950 0.5034 0.4950 0.7036
No log 11.0345 320 0.4870 0.6632 0.4870 0.6978
No log 11.1034 322 0.4889 0.6143 0.4889 0.6992
No log 11.1724 324 0.4850 0.6143 0.4850 0.6964
No log 11.2414 326 0.4786 0.6143 0.4786 0.6918
No log 11.3103 328 0.4965 0.5947 0.4965 0.7046
No log 11.3793 330 0.5357 0.5178 0.5357 0.7319
No log 11.4483 332 0.5482 0.5247 0.5482 0.7404
No log 11.5172 334 0.4896 0.5105 0.4896 0.6997
No log 11.5862 336 0.4890 0.4357 0.4890 0.6993
No log 11.6552 338 0.4881 0.4357 0.4881 0.6986
No log 11.7241 340 0.4870 0.4538 0.4870 0.6978
No log 11.7931 342 0.5218 0.5153 0.5218 0.7224
No log 11.8621 344 0.5815 0.5033 0.5815 0.7625
No log 11.9310 346 0.5563 0.5247 0.5563 0.7459
No log 12.0 348 0.5403 0.5538 0.5403 0.7350
No log 12.0690 350 0.5412 0.5736 0.5412 0.7356
No log 12.1379 352 0.5362 0.5662 0.5362 0.7323
No log 12.2069 354 0.5597 0.5256 0.5597 0.7481
No log 12.2759 356 0.5465 0.5662 0.5465 0.7393
No log 12.3448 358 0.5604 0.5025 0.5604 0.7486
No log 12.4138 360 0.5523 0.4497 0.5523 0.7432
No log 12.4828 362 0.5298 0.4547 0.5298 0.7279
No log 12.5517 364 0.5332 0.4964 0.5332 0.7302
No log 12.6207 366 0.5283 0.5472 0.5283 0.7268
No log 12.6897 368 0.5311 0.5501 0.5311 0.7288
No log 12.7586 370 0.5306 0.5813 0.5306 0.7284
No log 12.8276 372 0.5362 0.6388 0.5362 0.7322
No log 12.8966 374 0.5203 0.6009 0.5203 0.7213
No log 12.9655 376 0.5145 0.5886 0.5145 0.7173
No log 13.0345 378 0.5171 0.5166 0.5171 0.7191
No log 13.1034 380 0.5103 0.5698 0.5103 0.7144
No log 13.1724 382 0.5554 0.5283 0.5554 0.7453
No log 13.2414 384 0.5359 0.5265 0.5359 0.7320
No log 13.3103 386 0.5136 0.5550 0.5136 0.7167
No log 13.3793 388 0.5321 0.5324 0.5321 0.7294
No log 13.4483 390 0.5167 0.5455 0.5167 0.7188
No log 13.5172 392 0.5150 0.5640 0.5150 0.7176
No log 13.5862 394 0.5485 0.5554 0.5485 0.7406
No log 13.6552 396 0.5452 0.5383 0.5452 0.7384
No log 13.7241 398 0.5056 0.5719 0.5056 0.7111
No log 13.7931 400 0.5062 0.6129 0.5062 0.7115
No log 13.8621 402 0.5092 0.5567 0.5092 0.7136
No log 13.9310 404 0.5294 0.5373 0.5294 0.7276
No log 14.0 406 0.5264 0.5492 0.5264 0.7255
No log 14.0690 408 0.5363 0.5009 0.5363 0.7323
No log 14.1379 410 0.5268 0.4849 0.5268 0.7258
No log 14.2069 412 0.5139 0.4990 0.5139 0.7168
No log 14.2759 414 0.5114 0.4724 0.5114 0.7151
No log 14.3448 416 0.5114 0.3836 0.5114 0.7151
No log 14.4138 418 0.5176 0.4934 0.5176 0.7195
No log 14.4828 420 0.5338 0.4934 0.5338 0.7306
No log 14.5517 422 0.5248 0.4934 0.5248 0.7244
No log 14.6207 424 0.4982 0.5324 0.4982 0.7058
No log 14.6897 426 0.4906 0.5133 0.4906 0.7004
No log 14.7586 428 0.5026 0.5782 0.5026 0.7089
No log 14.8276 430 0.5115 0.6530 0.5115 0.7152
No log 14.8966 432 0.5320 0.6330 0.5320 0.7294
No log 14.9655 434 0.5784 0.5524 0.5784 0.7605
No log 15.0345 436 0.5573 0.5871 0.5573 0.7465
No log 15.1034 438 0.5496 0.5639 0.5496 0.7414
No log 15.1724 440 0.5195 0.5687 0.5195 0.7208
No log 15.2414 442 0.5254 0.5231 0.5254 0.7248
No log 15.3103 444 0.5277 0.4869 0.5277 0.7265
No log 15.3793 446 0.5090 0.4949 0.5090 0.7135
No log 15.4483 448 0.5251 0.4925 0.5251 0.7246
No log 15.5172 450 0.5566 0.5158 0.5566 0.7461
No log 15.5862 452 0.5463 0.5158 0.5463 0.7391
No log 15.6552 454 0.4966 0.5655 0.4966 0.7047
No log 15.7241 456 0.4939 0.5160 0.4939 0.7028
No log 15.7931 458 0.4948 0.5390 0.4948 0.7034
No log 15.8621 460 0.4808 0.6634 0.4808 0.6934
No log 15.9310 462 0.4826 0.6634 0.4826 0.6947
No log 16.0 464 0.5207 0.5881 0.5207 0.7216
No log 16.0690 466 0.5722 0.4825 0.5722 0.7564
No log 16.1379 468 0.5347 0.5230 0.5347 0.7312
No log 16.2069 470 0.4975 0.5731 0.4975 0.7053
No log 16.2759 472 0.4970 0.5715 0.4970 0.7050
No log 16.3448 474 0.5169 0.4502 0.5169 0.7190
No log 16.4138 476 0.5164 0.4753 0.5164 0.7186
No log 16.4828 478 0.5008 0.4768 0.5008 0.7077
No log 16.5517 480 0.5017 0.5089 0.5017 0.7083
No log 16.6207 482 0.5047 0.5533 0.5047 0.7104
No log 16.6897 484 0.5023 0.5753 0.5023 0.7088
No log 16.7586 486 0.5011 0.5405 0.5011 0.7079
No log 16.8276 488 0.5285 0.5687 0.5285 0.7270
No log 16.8966 490 0.5297 0.5479 0.5297 0.7278
No log 16.9655 492 0.4931 0.5373 0.4931 0.7022
No log 17.0345 494 0.4913 0.5437 0.4913 0.7009
No log 17.1034 496 0.5041 0.5195 0.5041 0.7100
No log 17.1724 498 0.4764 0.6408 0.4764 0.6902
0.2792 17.2414 500 0.5118 0.5432 0.5118 0.7154
0.2792 17.3103 502 0.6610 0.3847 0.6610 0.8130
0.2792 17.3793 504 0.7081 0.36 0.7081 0.8415
0.2792 17.4483 506 0.6466 0.4725 0.6466 0.8041
0.2792 17.5172 508 0.5616 0.5470 0.5616 0.7494
0.2792 17.5862 510 0.4972 0.6197 0.4972 0.7051
0.2792 17.6552 512 0.4901 0.5979 0.4901 0.7001
0.2792 17.7241 514 0.5067 0.6353 0.5067 0.7118
0.2792 17.7931 516 0.5054 0.6353 0.5054 0.7109
0.2792 17.8621 518 0.4921 0.6184 0.4921 0.7015
0.2792 17.9310 520 0.5234 0.5918 0.5234 0.7235
0.2792 18.0 522 0.5477 0.5438 0.5477 0.7401
0.2792 18.0690 524 0.5496 0.5603 0.5496 0.7413
0.2792 18.1379 526 0.5033 0.5495 0.5033 0.7095
0.2792 18.2069 528 0.4762 0.6555 0.4762 0.6901
0.2792 18.2759 530 0.4644 0.6242 0.4644 0.6814
0.2792 18.3448 532 0.4688 0.6908 0.4688 0.6847
0.2792 18.4138 534 0.4856 0.6491 0.4856 0.6969
0.2792 18.4828 536 0.5515 0.5636 0.5515 0.7427
0.2792 18.5517 538 0.5925 0.5249 0.5925 0.7698
0.2792 18.6207 540 0.6246 0.5129 0.6246 0.7903
0.2792 18.6897 542 0.5839 0.5112 0.5839 0.7641
0.2792 18.7586 544 0.5162 0.6061 0.5162 0.7185
0.2792 18.8276 546 0.4810 0.6156 0.4810 0.6936
0.2792 18.8966 548 0.4798 0.6111 0.4798 0.6927
0.2792 18.9655 550 0.4781 0.6242 0.4781 0.6915
0.2792 19.0345 552 0.4881 0.5463 0.4881 0.6986
0.2792 19.1034 554 0.5075 0.6013 0.5075 0.7124
0.2792 19.1724 556 0.5306 0.5161 0.5306 0.7284
0.2792 19.2414 558 0.5274 0.5569 0.5274 0.7262
0.2792 19.3103 560 0.5026 0.6492 0.5026 0.7090
0.2792 19.3793 562 0.5221 0.6158 0.5221 0.7226
0.2792 19.4483 564 0.5758 0.5236 0.5758 0.7588
0.2792 19.5172 566 0.5751 0.5363 0.5751 0.7584
0.2792 19.5862 568 0.5093 0.5841 0.5093 0.7136
0.2792 19.6552 570 0.4819 0.5875 0.4819 0.6942
0.2792 19.7241 572 0.4854 0.5379 0.4854 0.6967
0.2792 19.7931 574 0.5182 0.4821 0.5182 0.7198
0.2792 19.8621 576 0.5980 0.5263 0.5980 0.7733
0.2792 19.9310 578 0.6375 0.5051 0.6375 0.7984
0.2792 20.0 580 0.6133 0.5263 0.6133 0.7831

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k8_task7_organization

Finetuned
(4019)
this model