ArabicNewSplits6_FineTuningAraBERTFreeze_run2_AugV5_k9_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7442
  • Qwk: 0.4812
  • Mse: 0.7442
  • Rmse: 0.8627

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0870 2 6.4587 -0.0278 6.4587 2.5414
No log 0.1739 4 4.3438 -0.0299 4.3438 2.0842
No log 0.2609 6 3.0577 -0.0121 3.0578 1.7486
No log 0.3478 8 2.2140 0.0047 2.2140 1.4879
No log 0.4348 10 1.6057 -0.0270 1.6057 1.2672
No log 0.5217 12 1.1204 0.0148 1.1204 1.0585
No log 0.6087 14 0.9571 -0.0013 0.9571 0.9783
No log 0.6957 16 1.0344 -0.0356 1.0344 1.0170
No log 0.7826 18 1.2159 -0.0753 1.2159 1.1027
No log 0.8696 20 1.5580 -0.0632 1.5580 1.2482
No log 0.9565 22 1.4240 -0.0288 1.4240 1.1933
No log 1.0435 24 1.0403 -0.0472 1.0403 1.0200
No log 1.1304 26 0.7894 0.125 0.7894 0.8885
No log 1.2174 28 0.7173 0.2505 0.7173 0.8470
No log 1.3043 30 0.7174 0.2333 0.7174 0.8470
No log 1.3913 32 0.7330 0.1595 0.7330 0.8562
No log 1.4783 34 0.7682 0.1448 0.7682 0.8765
No log 1.5652 36 0.8219 0.1078 0.8219 0.9066
No log 1.6522 38 0.8932 0.1147 0.8932 0.9451
No log 1.7391 40 0.9477 0.0516 0.9477 0.9735
No log 1.8261 42 0.9532 0.1161 0.9532 0.9763
No log 1.9130 44 0.8812 0.1572 0.8812 0.9387
No log 2.0 46 0.8327 0.1907 0.8327 0.9125
No log 2.0870 48 0.7589 0.2757 0.7589 0.8711
No log 2.1739 50 0.7376 0.2683 0.7376 0.8589
No log 2.2609 52 0.7442 0.2917 0.7442 0.8627
No log 2.3478 54 0.7199 0.2844 0.7199 0.8485
No log 2.4348 56 0.6786 0.3052 0.6786 0.8237
No log 2.5217 58 0.6811 0.3595 0.6811 0.8253
No log 2.6087 60 0.6914 0.3740 0.6914 0.8315
No log 2.6957 62 0.7405 0.3581 0.7405 0.8605
No log 2.7826 64 0.7678 0.3111 0.7678 0.8763
No log 2.8696 66 0.7426 0.3434 0.7426 0.8617
No log 2.9565 68 0.6702 0.3241 0.6702 0.8187
No log 3.0435 70 0.6241 0.3622 0.6241 0.7900
No log 3.1304 72 0.6118 0.3740 0.6118 0.7822
No log 3.2174 74 0.5912 0.3845 0.5912 0.7689
No log 3.3043 76 0.5742 0.3898 0.5742 0.7577
No log 3.3913 78 0.5682 0.3898 0.5682 0.7538
No log 3.4783 80 0.5876 0.5012 0.5876 0.7665
No log 3.5652 82 0.6647 0.4700 0.6647 0.8153
No log 3.6522 84 0.7941 0.3706 0.7941 0.8911
No log 3.7391 86 0.7668 0.3766 0.7668 0.8757
No log 3.8261 88 0.6400 0.4922 0.6400 0.8000
No log 3.9130 90 0.5945 0.4938 0.5945 0.7710
No log 4.0 92 0.5739 0.4832 0.5739 0.7576
No log 4.0870 94 0.5757 0.4832 0.5757 0.7587
No log 4.1739 96 0.5978 0.4882 0.5978 0.7732
No log 4.2609 98 0.6381 0.4491 0.6381 0.7988
No log 4.3478 100 0.6423 0.4491 0.6423 0.8014
No log 4.4348 102 0.5895 0.4700 0.5895 0.7678
No log 4.5217 104 0.5722 0.4974 0.5722 0.7565
No log 4.6087 106 0.5622 0.5027 0.5622 0.7498
No log 4.6957 108 0.5630 0.5015 0.5630 0.7504
No log 4.7826 110 0.5616 0.5027 0.5616 0.7494
No log 4.8696 112 0.5627 0.5071 0.5627 0.7501
No log 4.9565 114 0.6170 0.4424 0.6170 0.7855
No log 5.0435 116 0.6790 0.4628 0.6790 0.8240
No log 5.1304 118 0.6907 0.4648 0.6907 0.8311
No log 5.2174 120 0.6153 0.4381 0.6153 0.7844
No log 5.3043 122 0.5729 0.4851 0.5729 0.7569
No log 5.3913 124 0.5545 0.4713 0.5545 0.7447
No log 5.4783 126 0.5428 0.5030 0.5428 0.7368
No log 5.5652 128 0.5399 0.5030 0.5399 0.7348
No log 5.6522 130 0.5374 0.4757 0.5374 0.7331
No log 5.7391 132 0.5677 0.5250 0.5677 0.7534
No log 5.8261 134 0.6548 0.5015 0.6548 0.8092
No log 5.9130 136 0.7124 0.4867 0.7124 0.8440
No log 6.0 138 0.7365 0.4752 0.7365 0.8582
No log 6.0870 140 0.6935 0.4779 0.6935 0.8327
No log 6.1739 142 0.6552 0.5058 0.6552 0.8095
No log 6.2609 144 0.6552 0.4871 0.6552 0.8094
No log 6.3478 146 0.5856 0.5522 0.5856 0.7652
No log 6.4348 148 0.5628 0.5315 0.5628 0.7502
No log 6.5217 150 0.5781 0.5430 0.5781 0.7603
No log 6.6087 152 0.6040 0.5103 0.6040 0.7771
No log 6.6957 154 0.6061 0.5156 0.6061 0.7785
No log 6.7826 156 0.6627 0.5211 0.6627 0.8141
No log 6.8696 158 0.7258 0.4868 0.7258 0.8519
No log 6.9565 160 0.6922 0.5216 0.6922 0.8320
No log 7.0435 162 0.6708 0.5626 0.6708 0.8190
No log 7.1304 164 0.6436 0.5461 0.6436 0.8023
No log 7.2174 166 0.6439 0.5493 0.6439 0.8024
No log 7.3043 168 0.6742 0.5256 0.6742 0.8211
No log 7.3913 170 0.7324 0.5403 0.7324 0.8558
No log 7.4783 172 0.7115 0.5331 0.7115 0.8435
No log 7.5652 174 0.6350 0.5360 0.6350 0.7969
No log 7.6522 176 0.6030 0.54 0.6030 0.7765
No log 7.7391 178 0.6021 0.5465 0.6021 0.7759
No log 7.8261 180 0.5940 0.5331 0.5940 0.7707
No log 7.9130 182 0.5922 0.5636 0.5922 0.7695
No log 8.0 184 0.6024 0.5759 0.6024 0.7761
No log 8.0870 186 0.6109 0.5235 0.6109 0.7816
No log 8.1739 188 0.6318 0.5258 0.6318 0.7948
No log 8.2609 190 0.6747 0.5067 0.6747 0.8214
No log 8.3478 192 0.7046 0.5096 0.7046 0.8394
No log 8.4348 194 0.7212 0.4827 0.7212 0.8492
No log 8.5217 196 0.6971 0.5022 0.6971 0.8349
No log 8.6087 198 0.6956 0.5421 0.6956 0.8340
No log 8.6957 200 0.7247 0.5021 0.7247 0.8513
No log 8.7826 202 0.7936 0.4967 0.7936 0.8909
No log 8.8696 204 0.8295 0.5056 0.8295 0.9108
No log 8.9565 206 0.7804 0.4877 0.7804 0.8834
No log 9.0435 208 0.7621 0.4557 0.7621 0.8730
No log 9.1304 210 0.7245 0.4764 0.7245 0.8512
No log 9.2174 212 0.7061 0.5488 0.7061 0.8403
No log 9.3043 214 0.7034 0.5211 0.7034 0.8387
No log 9.3913 216 0.7315 0.4758 0.7315 0.8553
No log 9.4783 218 0.7541 0.4755 0.7541 0.8684
No log 9.5652 220 0.7336 0.4625 0.7336 0.8565
No log 9.6522 222 0.7046 0.5277 0.7046 0.8394
No log 9.7391 224 0.7182 0.5298 0.7182 0.8475
No log 9.8261 226 0.7344 0.5181 0.7344 0.8570
No log 9.9130 228 0.7217 0.5331 0.7217 0.8495
No log 10.0 230 0.7471 0.4758 0.7471 0.8644
No log 10.0870 232 0.8181 0.4995 0.8181 0.9045
No log 10.1739 234 0.8032 0.4979 0.8032 0.8962
No log 10.2609 236 0.7492 0.5053 0.7492 0.8656
No log 10.3478 238 0.7454 0.4936 0.7454 0.8633
No log 10.4348 240 0.7519 0.4895 0.7519 0.8671
No log 10.5217 242 0.7587 0.4682 0.7587 0.8710
No log 10.6087 244 0.7697 0.4802 0.7697 0.8773
No log 10.6957 246 0.7745 0.4862 0.7745 0.8800
No log 10.7826 248 0.7816 0.4626 0.7816 0.8841
No log 10.8696 250 0.7757 0.4536 0.7757 0.8808
No log 10.9565 252 0.7464 0.4715 0.7464 0.8639
No log 11.0435 254 0.7404 0.5286 0.7404 0.8605
No log 11.1304 256 0.7488 0.5261 0.7488 0.8653
No log 11.2174 258 0.7460 0.5282 0.7460 0.8637
No log 11.3043 260 0.7460 0.4581 0.7460 0.8637
No log 11.3913 262 0.7543 0.4942 0.7543 0.8685
No log 11.4783 264 0.7725 0.4997 0.7725 0.8789
No log 11.5652 266 0.7898 0.4857 0.7898 0.8887
No log 11.6522 268 0.7989 0.5056 0.7989 0.8938
No log 11.7391 270 0.8043 0.5256 0.8043 0.8968
No log 11.8261 272 0.7872 0.5290 0.7872 0.8873
No log 11.9130 274 0.7831 0.5052 0.7831 0.8849
No log 12.0 276 0.7649 0.5033 0.7649 0.8746
No log 12.0870 278 0.8020 0.4686 0.8020 0.8956
No log 12.1739 280 0.8167 0.4539 0.8167 0.9037
No log 12.2609 282 0.8092 0.4864 0.8092 0.8995
No log 12.3478 284 0.8143 0.4942 0.8143 0.9024
No log 12.4348 286 0.7466 0.4688 0.7466 0.8641
No log 12.5217 288 0.7144 0.5 0.7144 0.8452
No log 12.6087 290 0.7019 0.5 0.7019 0.8378
No log 12.6957 292 0.7128 0.4724 0.7128 0.8443
No log 12.7826 294 0.7292 0.4709 0.7292 0.8539
No log 12.8696 296 0.7326 0.4850 0.7326 0.8559
No log 12.9565 298 0.7043 0.4716 0.7043 0.8392
No log 13.0435 300 0.7098 0.5403 0.7098 0.8425
No log 13.1304 302 0.7140 0.5250 0.7140 0.8450
No log 13.2174 304 0.7235 0.5175 0.7235 0.8506
No log 13.3043 306 0.7263 0.5129 0.7263 0.8522
No log 13.3913 308 0.7569 0.4341 0.7569 0.8700
No log 13.4783 310 0.7650 0.4847 0.7650 0.8746
No log 13.5652 312 0.7278 0.4755 0.7278 0.8531
No log 13.6522 314 0.7279 0.5038 0.7279 0.8531
No log 13.7391 316 0.7351 0.4935 0.7351 0.8574
No log 13.8261 318 0.7026 0.5089 0.7026 0.8382
No log 13.9130 320 0.6831 0.5206 0.6831 0.8265
No log 14.0 322 0.7346 0.5025 0.7346 0.8571
No log 14.0870 324 0.7325 0.4966 0.7325 0.8559
No log 14.1739 326 0.7004 0.4725 0.7004 0.8369
No log 14.2609 328 0.7027 0.4700 0.7027 0.8383
No log 14.3478 330 0.7361 0.4419 0.7361 0.8580
No log 14.4348 332 0.7641 0.4602 0.7641 0.8741
No log 14.5217 334 0.7603 0.4725 0.7603 0.8720
No log 14.6087 336 0.7066 0.3924 0.7066 0.8406
No log 14.6957 338 0.6929 0.3896 0.6929 0.8324
No log 14.7826 340 0.7118 0.4506 0.7118 0.8437
No log 14.8696 342 0.7056 0.4440 0.7056 0.8400
No log 14.9565 344 0.6805 0.4481 0.6805 0.8249
No log 15.0435 346 0.6578 0.4793 0.6578 0.8111
No log 15.1304 348 0.6456 0.4854 0.6456 0.8035
No log 15.2174 350 0.6520 0.4978 0.6520 0.8074
No log 15.3043 352 0.6556 0.4340 0.6556 0.8097
No log 15.3913 354 0.6715 0.4655 0.6715 0.8195
No log 15.4783 356 0.6838 0.4655 0.6838 0.8269
No log 15.5652 358 0.7036 0.4709 0.7036 0.8388
No log 15.6522 360 0.7090 0.4880 0.7090 0.8420
No log 15.7391 362 0.6971 0.4337 0.6971 0.8349
No log 15.8261 364 0.6938 0.4798 0.6938 0.8330
No log 15.9130 366 0.6969 0.4915 0.6969 0.8348
No log 16.0 368 0.6980 0.5205 0.6980 0.8355
No log 16.0870 370 0.7045 0.5386 0.7045 0.8393
No log 16.1739 372 0.6978 0.4915 0.6978 0.8353
No log 16.2609 374 0.7172 0.5006 0.7172 0.8469
No log 16.3478 376 0.7172 0.5060 0.7172 0.8469
No log 16.4348 378 0.7230 0.4971 0.7230 0.8503
No log 16.5217 380 0.7363 0.4901 0.7363 0.8581
No log 16.6087 382 0.7621 0.4901 0.7621 0.8730
No log 16.6957 384 0.7453 0.5017 0.7453 0.8633
No log 16.7826 386 0.7197 0.4622 0.7197 0.8483
No log 16.8696 388 0.7154 0.4937 0.7154 0.8458
No log 16.9565 390 0.7079 0.4590 0.7079 0.8413
No log 17.0435 392 0.7154 0.4916 0.7154 0.8458
No log 17.1304 394 0.7204 0.5142 0.7204 0.8487
No log 17.2174 396 0.7102 0.4916 0.7102 0.8427
No log 17.3043 398 0.6994 0.5126 0.6994 0.8363
No log 17.3913 400 0.7192 0.5280 0.7192 0.8481
No log 17.4783 402 0.7493 0.4764 0.7493 0.8656
No log 17.5652 404 0.7420 0.4839 0.7420 0.8614
No log 17.6522 406 0.7292 0.5058 0.7292 0.8539
No log 17.7391 408 0.7382 0.4964 0.7382 0.8592
No log 17.8261 410 0.7279 0.4963 0.7279 0.8532
No log 17.9130 412 0.7526 0.4608 0.7526 0.8675
No log 18.0 414 0.7628 0.4627 0.7628 0.8734
No log 18.0870 416 0.7443 0.5071 0.7443 0.8627
No log 18.1739 418 0.7645 0.5052 0.7645 0.8743
No log 18.2609 420 0.7721 0.5122 0.7721 0.8787
No log 18.3478 422 0.7536 0.4942 0.7536 0.8681
No log 18.4348 424 0.7609 0.5463 0.7609 0.8723
No log 18.5217 426 0.7512 0.5448 0.7512 0.8667
No log 18.6087 428 0.7161 0.5097 0.7161 0.8462
No log 18.6957 430 0.7070 0.4940 0.7070 0.8408
No log 18.7826 432 0.7306 0.5270 0.7306 0.8547
No log 18.8696 434 0.7126 0.5281 0.7126 0.8442
No log 18.9565 436 0.6780 0.5044 0.6780 0.8234
No log 19.0435 438 0.6995 0.5211 0.6995 0.8364
No log 19.1304 440 0.7116 0.5211 0.7116 0.8436
No log 19.2174 442 0.7019 0.5225 0.7019 0.8378
No log 19.3043 444 0.7305 0.4928 0.7305 0.8547
No log 19.3913 446 0.7900 0.5188 0.7900 0.8888
No log 19.4783 448 0.7808 0.4858 0.7808 0.8836
No log 19.5652 450 0.7516 0.4940 0.7516 0.8669
No log 19.6522 452 0.7533 0.5113 0.7533 0.8679
No log 19.7391 454 0.7498 0.4958 0.7498 0.8659
No log 19.8261 456 0.7428 0.4958 0.7428 0.8619
No log 19.9130 458 0.7235 0.5113 0.7235 0.8506
No log 20.0 460 0.7058 0.5116 0.7058 0.8401
No log 20.0870 462 0.6878 0.4889 0.6878 0.8293
No log 20.1739 464 0.6740 0.4856 0.6740 0.8210
No log 20.2609 466 0.6753 0.5018 0.6753 0.8218
No log 20.3478 468 0.6827 0.5068 0.6827 0.8263
No log 20.4348 470 0.6937 0.5358 0.6937 0.8329
No log 20.5217 472 0.7074 0.4791 0.7074 0.8411
No log 20.6087 474 0.7377 0.4979 0.7377 0.8589
No log 20.6957 476 0.7620 0.4979 0.7620 0.8729
No log 20.7826 478 0.7663 0.5039 0.7663 0.8754
No log 20.8696 480 0.7741 0.4726 0.7741 0.8798
No log 20.9565 482 0.7667 0.4959 0.7667 0.8756
No log 21.0435 484 0.7532 0.5145 0.7532 0.8679
No log 21.1304 486 0.7387 0.5147 0.7387 0.8595
No log 21.2174 488 0.7023 0.4983 0.7023 0.8380
No log 21.3043 490 0.6812 0.5227 0.6812 0.8253
No log 21.3913 492 0.7159 0.4840 0.7159 0.8461
No log 21.4783 494 0.7204 0.4986 0.7204 0.8488
No log 21.5652 496 0.6797 0.4725 0.6797 0.8244
No log 21.6522 498 0.6906 0.5084 0.6906 0.8310
0.546 21.7391 500 0.7346 0.4864 0.7346 0.8571
0.546 21.8261 502 0.7494 0.4906 0.7494 0.8657
0.546 21.9130 504 0.7088 0.4997 0.7088 0.8419
0.546 22.0 506 0.7016 0.4902 0.7016 0.8376
0.546 22.0870 508 0.7336 0.4909 0.7336 0.8565
0.546 22.1739 510 0.7248 0.5018 0.7248 0.8514
0.546 22.2609 512 0.7033 0.4887 0.7033 0.8386
0.546 22.3478 514 0.7186 0.5015 0.7186 0.8477
0.546 22.4348 516 0.7521 0.4868 0.7521 0.8672
0.546 22.5217 518 0.7459 0.5047 0.7459 0.8637
0.546 22.6087 520 0.7260 0.5017 0.7260 0.8521
0.546 22.6957 522 0.7565 0.4865 0.7565 0.8698
0.546 22.7826 524 0.8201 0.5094 0.8201 0.9056
0.546 22.8696 526 0.8498 0.4926 0.8498 0.9218
0.546 22.9565 528 0.7903 0.4827 0.7903 0.8890
0.546 23.0435 530 0.7438 0.5176 0.7438 0.8624
0.546 23.1304 532 0.7819 0.4864 0.7819 0.8843
0.546 23.2174 534 0.8063 0.4773 0.8063 0.8979
0.546 23.3043 536 0.7924 0.4984 0.7924 0.8902
0.546 23.3913 538 0.7899 0.5099 0.7899 0.8888
0.546 23.4783 540 0.7922 0.5047 0.7922 0.8901
0.546 23.5652 542 0.7956 0.4956 0.7956 0.8920
0.546 23.6522 544 0.8076 0.4833 0.8076 0.8987
0.546 23.7391 546 0.8237 0.4972 0.8237 0.9076
0.546 23.8261 548 0.8126 0.4875 0.8126 0.9015
0.546 23.9130 550 0.7826 0.4800 0.7826 0.8846
0.546 24.0 552 0.7548 0.5169 0.7548 0.8688
0.546 24.0870 554 0.7470 0.5126 0.7470 0.8643
0.546 24.1739 556 0.7487 0.5075 0.7487 0.8653
0.546 24.2609 558 0.7907 0.4733 0.7907 0.8892
0.546 24.3478 560 0.8080 0.4785 0.8080 0.8989
0.546 24.4348 562 0.7835 0.5104 0.7835 0.8852
0.546 24.5217 564 0.7724 0.4909 0.7724 0.8789
0.546 24.6087 566 0.7682 0.4604 0.7682 0.8765
0.546 24.6957 568 0.7650 0.4684 0.7650 0.8746
0.546 24.7826 570 0.7594 0.4615 0.7594 0.8714
0.546 24.8696 572 0.7727 0.4867 0.7727 0.8791
0.546 24.9565 574 0.7693 0.4725 0.7693 0.8771
0.546 25.0435 576 0.7475 0.4724 0.7475 0.8646
0.546 25.1304 578 0.7548 0.4590 0.7548 0.8688
0.546 25.2174 580 0.7658 0.4660 0.7658 0.8751
0.546 25.3043 582 0.7442 0.4812 0.7442 0.8627

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run2_AugV5_k9_task2_organization

Finetuned
(4023)
this model