ArabicNewSplits6_FineTuningAraBERTFreeze_run3_AugV5_k18_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7623
  • Qwk: 0.4854
  • Mse: 0.7623
  • Rmse: 0.8731

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0444 2 6.5198 -0.0278 6.5198 2.5534
No log 0.0889 4 4.4223 -0.0312 4.4223 2.1029
No log 0.1333 6 3.0763 -0.0192 3.0763 1.7539
No log 0.1778 8 2.2252 0.0011 2.2252 1.4917
No log 0.2222 10 1.6529 -0.0397 1.6529 1.2857
No log 0.2667 12 1.4944 -0.0777 1.4944 1.2224
No log 0.3111 14 1.5530 -0.1249 1.5530 1.2462
No log 0.3556 16 1.4639 -0.0835 1.4639 1.2099
No log 0.4 18 1.2132 0.0148 1.2132 1.1015
No log 0.4444 20 1.1423 -0.0130 1.1423 1.0688
No log 0.4889 22 1.1678 0.0234 1.1678 1.0807
No log 0.5333 24 1.0748 0.0066 1.0748 1.0367
No log 0.5778 26 0.9353 0.0128 0.9353 0.9671
No log 0.6222 28 0.7806 0.2203 0.7806 0.8835
No log 0.6667 30 0.7381 0.2291 0.7381 0.8591
No log 0.7111 32 0.7302 0.1957 0.7302 0.8545
No log 0.7556 34 0.7332 0.2083 0.7332 0.8563
No log 0.8 36 0.7524 0.0969 0.7524 0.8674
No log 0.8444 38 0.7681 0.1236 0.7681 0.8764
No log 0.8889 40 0.8044 0.1626 0.8044 0.8969
No log 0.9333 42 0.9008 0.1947 0.9008 0.9491
No log 0.9778 44 0.9795 0.2203 0.9795 0.9897
No log 1.0222 46 1.0870 0.1839 1.0870 1.0426
No log 1.0667 48 1.2987 0.1907 1.2987 1.1396
No log 1.1111 50 1.3472 0.1860 1.3472 1.1607
No log 1.1556 52 1.1638 0.2080 1.1638 1.0788
No log 1.2 54 0.9829 0.3043 0.9829 0.9914
No log 1.2444 56 0.8913 0.3227 0.8913 0.9441
No log 1.2889 58 0.7733 0.3185 0.7733 0.8794
No log 1.3333 60 0.7322 0.4159 0.7322 0.8557
No log 1.3778 62 0.6898 0.3932 0.6898 0.8305
No log 1.4222 64 0.6821 0.3938 0.6821 0.8259
No log 1.4667 66 0.7015 0.4004 0.7015 0.8376
No log 1.5111 68 0.7614 0.4122 0.7614 0.8726
No log 1.5556 70 0.7937 0.3744 0.7937 0.8909
No log 1.6 72 0.7883 0.3982 0.7883 0.8878
No log 1.6444 74 0.7353 0.3782 0.7353 0.8575
No log 1.6889 76 0.7258 0.3751 0.7258 0.8520
No log 1.7333 78 0.7344 0.3927 0.7344 0.8570
No log 1.7778 80 0.6826 0.4015 0.6826 0.8262
No log 1.8222 82 0.6730 0.3922 0.6730 0.8203
No log 1.8667 84 0.6705 0.4204 0.6705 0.8188
No log 1.9111 86 0.6769 0.4099 0.6769 0.8227
No log 1.9556 88 0.6514 0.4572 0.6514 0.8071
No log 2.0 90 0.6415 0.4559 0.6415 0.8010
No log 2.0444 92 0.6525 0.4628 0.6525 0.8078
No log 2.0889 94 0.6700 0.4567 0.6700 0.8186
No log 2.1333 96 0.7575 0.4477 0.7575 0.8703
No log 2.1778 98 0.8310 0.4260 0.8310 0.9116
No log 2.2222 100 0.7583 0.4494 0.7583 0.8708
No log 2.2667 102 0.6506 0.4831 0.6506 0.8066
No log 2.3111 104 0.6106 0.5041 0.6106 0.7814
No log 2.3556 106 0.5999 0.4765 0.5999 0.7745
No log 2.4 108 0.6064 0.4865 0.6064 0.7787
No log 2.4444 110 0.6177 0.5 0.6177 0.7859
No log 2.4889 112 0.6329 0.5258 0.6329 0.7956
No log 2.5333 114 0.6282 0.5288 0.6282 0.7926
No log 2.5778 116 0.6221 0.5257 0.6221 0.7887
No log 2.6222 118 0.6198 0.5069 0.6198 0.7873
No log 2.6667 120 0.6258 0.4964 0.6258 0.7911
No log 2.7111 122 0.6303 0.5314 0.6303 0.7939
No log 2.7556 124 0.6431 0.5075 0.6431 0.8019
No log 2.8 126 0.6580 0.5019 0.6580 0.8112
No log 2.8444 128 0.6428 0.4990 0.6428 0.8017
No log 2.8889 130 0.6598 0.4809 0.6598 0.8123
No log 2.9333 132 0.7063 0.4626 0.7063 0.8404
No log 2.9778 134 0.6914 0.4685 0.6914 0.8315
No log 3.0222 136 0.6386 0.4796 0.6386 0.7991
No log 3.0667 138 0.6165 0.5212 0.6165 0.7852
No log 3.1111 140 0.6205 0.5212 0.6205 0.7877
No log 3.1556 142 0.6147 0.5446 0.6147 0.7840
No log 3.2 144 0.6092 0.5354 0.6092 0.7805
No log 3.2444 146 0.6235 0.5429 0.6235 0.7896
No log 3.2889 148 0.6483 0.5271 0.6483 0.8052
No log 3.3333 150 0.7008 0.4287 0.7008 0.8371
No log 3.3778 152 0.7263 0.4476 0.7263 0.8523
No log 3.4222 154 0.7134 0.4530 0.7134 0.8446
No log 3.4667 156 0.7220 0.4469 0.7220 0.8497
No log 3.5111 158 0.7061 0.4949 0.7061 0.8403
No log 3.5556 160 0.7123 0.494 0.7123 0.8440
No log 3.6 162 0.7243 0.5027 0.7243 0.8511
No log 3.6444 164 0.7443 0.5057 0.7443 0.8627
No log 3.6889 166 0.7447 0.5102 0.7447 0.8630
No log 3.7333 168 0.7661 0.4708 0.7661 0.8753
No log 3.7778 170 0.7616 0.4536 0.7616 0.8727
No log 3.8222 172 0.7173 0.4836 0.7173 0.8470
No log 3.8667 174 0.6476 0.4933 0.6476 0.8047
No log 3.9111 176 0.6191 0.5126 0.6191 0.7869
No log 3.9556 178 0.6188 0.5130 0.6188 0.7866
No log 4.0 180 0.6129 0.5218 0.6129 0.7829
No log 4.0444 182 0.6065 0.4975 0.6065 0.7788
No log 4.0889 184 0.6179 0.5059 0.6179 0.7860
No log 4.1333 186 0.6340 0.5144 0.6340 0.7962
No log 4.1778 188 0.6487 0.5309 0.6487 0.8054
No log 4.2222 190 0.6642 0.5229 0.6642 0.8150
No log 4.2667 192 0.6807 0.5022 0.6807 0.8250
No log 4.3111 194 0.7014 0.5044 0.7014 0.8375
No log 4.3556 196 0.7118 0.5043 0.7118 0.8437
No log 4.4 198 0.7133 0.5021 0.7133 0.8445
No log 4.4444 200 0.7104 0.4889 0.7104 0.8429
No log 4.4889 202 0.7108 0.4688 0.7108 0.8431
No log 4.5333 204 0.7158 0.4688 0.7158 0.8460
No log 4.5778 206 0.7177 0.4897 0.7177 0.8472
No log 4.6222 208 0.7157 0.4671 0.7157 0.8460
No log 4.6667 210 0.7047 0.4573 0.7047 0.8395
No log 4.7111 212 0.7118 0.4586 0.7118 0.8437
No log 4.7556 214 0.7105 0.4586 0.7105 0.8429
No log 4.8 216 0.7108 0.4586 0.7108 0.8431
No log 4.8444 218 0.7149 0.4586 0.7149 0.8455
No log 4.8889 220 0.7213 0.4437 0.7213 0.8493
No log 4.9333 222 0.7277 0.4839 0.7277 0.8530
No log 4.9778 224 0.7452 0.5132 0.7452 0.8633
No log 5.0222 226 0.7560 0.5178 0.7560 0.8695
No log 5.0667 228 0.7484 0.4961 0.7484 0.8651
No log 5.1111 230 0.7918 0.4636 0.7918 0.8898
No log 5.1556 232 0.8416 0.4285 0.8416 0.9174
No log 5.2 234 0.8041 0.4665 0.8041 0.8967
No log 5.2444 236 0.7324 0.4966 0.7324 0.8558
No log 5.2889 238 0.7206 0.5351 0.7206 0.8489
No log 5.3333 240 0.7176 0.5211 0.7176 0.8471
No log 5.3778 242 0.7233 0.5088 0.7233 0.8505
No log 5.4222 244 0.7445 0.5382 0.7445 0.8628
No log 5.4667 246 0.7654 0.5109 0.7654 0.8749
No log 5.5111 248 0.7950 0.5105 0.7950 0.8916
No log 5.5556 250 0.8115 0.5224 0.8115 0.9008
No log 5.6 252 0.8223 0.5218 0.8223 0.9068
No log 5.6444 254 0.8266 0.5145 0.8266 0.9092
No log 5.6889 256 0.8228 0.5108 0.8228 0.9071
No log 5.7333 258 0.8354 0.5005 0.8354 0.9140
No log 5.7778 260 0.8332 0.4997 0.8332 0.9128
No log 5.8222 262 0.8461 0.4929 0.8461 0.9198
No log 5.8667 264 0.8118 0.5095 0.8118 0.9010
No log 5.9111 266 0.7888 0.4945 0.7888 0.8881
No log 5.9556 268 0.7700 0.5143 0.7700 0.8775
No log 6.0 270 0.7746 0.4934 0.7746 0.8801
No log 6.0444 272 0.7890 0.5030 0.7890 0.8883
No log 6.0889 274 0.7884 0.5119 0.7884 0.8879
No log 6.1333 276 0.7864 0.4776 0.7864 0.8868
No log 6.1778 278 0.7957 0.4820 0.7957 0.8920
No log 6.2222 280 0.7868 0.4566 0.7868 0.8870
No log 6.2667 282 0.7643 0.4330 0.7643 0.8742
No log 6.3111 284 0.7487 0.4859 0.7487 0.8653
No log 6.3556 286 0.8133 0.4890 0.8133 0.9018
No log 6.4 288 0.8385 0.4510 0.8385 0.9157
No log 6.4444 290 0.8000 0.4890 0.8000 0.8945
No log 6.4889 292 0.7599 0.5360 0.7599 0.8717
No log 6.5333 294 0.7622 0.4648 0.7622 0.8731
No log 6.5778 296 0.7878 0.4994 0.7878 0.8876
No log 6.6222 298 0.7636 0.4975 0.7636 0.8738
No log 6.6667 300 0.7729 0.525 0.7729 0.8791
No log 6.7111 302 0.8818 0.4311 0.8818 0.9390
No log 6.7556 304 0.8800 0.4312 0.8800 0.9381
No log 6.8 306 0.7890 0.4790 0.7890 0.8883
No log 6.8444 308 0.7618 0.4984 0.7618 0.8728
No log 6.8889 310 0.7634 0.5071 0.7634 0.8737
No log 6.9333 312 0.7584 0.5147 0.7584 0.8708
No log 6.9778 314 0.7859 0.5201 0.7859 0.8865
No log 7.0222 316 0.7981 0.4930 0.7981 0.8933
No log 7.0667 318 0.8488 0.4472 0.8488 0.9213
No log 7.1111 320 0.8814 0.4421 0.8814 0.9388
No log 7.1556 322 0.8358 0.4930 0.8358 0.9142
No log 7.2 324 0.7787 0.4847 0.7787 0.8824
No log 7.2444 326 0.7617 0.5324 0.7617 0.8728
No log 7.2889 328 0.7606 0.5324 0.7606 0.8721
No log 7.3333 330 0.7855 0.4905 0.7855 0.8863
No log 7.3778 332 0.8864 0.4529 0.8864 0.9415
No log 7.4222 334 0.9170 0.4371 0.9170 0.9576
No log 7.4667 336 0.8627 0.4497 0.8627 0.9288
No log 7.5111 338 0.8152 0.4605 0.8152 0.9029
No log 7.5556 340 0.8207 0.4698 0.8207 0.9059
No log 7.6 342 0.8356 0.5071 0.8356 0.9141
No log 7.6444 344 0.8321 0.4881 0.8321 0.9122
No log 7.6889 346 0.8205 0.4639 0.8205 0.9058
No log 7.7333 348 0.8428 0.4948 0.8428 0.9180
No log 7.7778 350 0.8285 0.5146 0.8285 0.9102
No log 7.8222 352 0.8167 0.5303 0.8167 0.9037
No log 7.8667 354 0.8070 0.5049 0.8070 0.8983
No log 7.9111 356 0.7800 0.5292 0.7800 0.8832
No log 7.9556 358 0.7489 0.5081 0.7489 0.8654
No log 8.0 360 0.7423 0.5176 0.7423 0.8615
No log 8.0444 362 0.7585 0.5635 0.7585 0.8709
No log 8.0889 364 0.7793 0.5138 0.7793 0.8828
No log 8.1333 366 0.8081 0.4592 0.8081 0.8989
No log 8.1778 368 0.8131 0.4997 0.8131 0.9017
No log 8.2222 370 0.8186 0.4717 0.8186 0.9048
No log 8.2667 372 0.8419 0.4662 0.8419 0.9175
No log 8.3111 374 0.8591 0.4572 0.8591 0.9269
No log 8.3556 376 0.8571 0.4494 0.8571 0.9258
No log 8.4 378 0.8595 0.4699 0.8595 0.9271
No log 8.4444 380 0.8360 0.4508 0.8360 0.9143
No log 8.4889 382 0.8153 0.4480 0.8153 0.9029
No log 8.5333 384 0.7980 0.4382 0.7980 0.8933
No log 8.5778 386 0.8046 0.4538 0.8046 0.8970
No log 8.6222 388 0.8143 0.4174 0.8143 0.9024
No log 8.6667 390 0.7811 0.4161 0.7811 0.8838
No log 8.7111 392 0.7397 0.4549 0.7397 0.8601
No log 8.7556 394 0.7271 0.4853 0.7271 0.8527
No log 8.8 396 0.7272 0.4657 0.7272 0.8528
No log 8.8444 398 0.7310 0.4841 0.7310 0.8550
No log 8.8889 400 0.7377 0.4779 0.7377 0.8589
No log 8.9333 402 0.7549 0.4689 0.7549 0.8688
No log 8.9778 404 0.7821 0.4873 0.7821 0.8844
No log 9.0222 406 0.7908 0.5340 0.7908 0.8892
No log 9.0667 408 0.8014 0.4503 0.8014 0.8952
No log 9.1111 410 0.8199 0.4699 0.8199 0.9055
No log 9.1556 412 0.8125 0.4784 0.8125 0.9014
No log 9.2 414 0.8021 0.5417 0.8021 0.8956
No log 9.2444 416 0.7884 0.5446 0.7884 0.8879
No log 9.2889 418 0.7681 0.5431 0.7681 0.8764
No log 9.3333 420 0.7708 0.5292 0.7708 0.8779
No log 9.3778 422 0.7755 0.4748 0.7755 0.8806
No log 9.4222 424 0.7869 0.4617 0.7869 0.8871
No log 9.4667 426 0.7812 0.4992 0.7812 0.8838
No log 9.5111 428 0.7719 0.5 0.7719 0.8786
No log 9.5556 430 0.7733 0.5346 0.7733 0.8794
No log 9.6 432 0.7734 0.4767 0.7734 0.8794
No log 9.6444 434 0.7883 0.4987 0.7883 0.8879
No log 9.6889 436 0.7740 0.4636 0.7740 0.8798
No log 9.7333 438 0.7861 0.4770 0.7861 0.8866
No log 9.7778 440 0.8251 0.5155 0.8251 0.9084
No log 9.8222 442 0.8490 0.4975 0.8490 0.9214
No log 9.8667 444 0.8487 0.4765 0.8487 0.9212
No log 9.9111 446 0.8758 0.4524 0.8758 0.9358
No log 9.9556 448 0.9227 0.4874 0.9227 0.9606
No log 10.0 450 0.9363 0.4802 0.9363 0.9676
No log 10.0444 452 0.9030 0.4738 0.9030 0.9503
No log 10.0889 454 0.8565 0.4719 0.8565 0.9255
No log 10.1333 456 0.8427 0.5033 0.8427 0.9180
No log 10.1778 458 0.8173 0.5226 0.8173 0.9040
No log 10.2222 460 0.7851 0.5231 0.7851 0.8861
No log 10.2667 462 0.7648 0.5149 0.7648 0.8745
No log 10.3111 464 0.7719 0.5143 0.7719 0.8786
No log 10.3556 466 0.7440 0.5241 0.7440 0.8626
No log 10.4 468 0.7123 0.5064 0.7123 0.8440
No log 10.4444 470 0.7133 0.5143 0.7133 0.8445
No log 10.4889 472 0.7342 0.5149 0.7342 0.8568
No log 10.5333 474 0.7625 0.5079 0.7625 0.8732
No log 10.5778 476 0.7856 0.5008 0.7856 0.8864
No log 10.6222 478 0.8076 0.5172 0.8076 0.8987
No log 10.6667 480 0.8179 0.4964 0.8179 0.9044
No log 10.7111 482 0.8250 0.4898 0.8250 0.9083
No log 10.7556 484 0.8296 0.4913 0.8296 0.9108
No log 10.8 486 0.8270 0.4906 0.8270 0.9094
No log 10.8444 488 0.8263 0.5066 0.8263 0.9090
No log 10.8889 490 0.8240 0.4898 0.8240 0.9077
No log 10.9333 492 0.8041 0.4767 0.8041 0.8967
No log 10.9778 494 0.8149 0.5032 0.8149 0.9027
No log 11.0222 496 0.8165 0.5087 0.8165 0.9036
No log 11.0667 498 0.7995 0.4883 0.7995 0.8942
0.5394 11.1111 500 0.8104 0.5045 0.8104 0.9002
0.5394 11.1556 502 0.8348 0.4626 0.8348 0.9137
0.5394 11.2 504 0.8340 0.5008 0.8340 0.9133
0.5394 11.2444 506 0.8158 0.5045 0.8158 0.9032
0.5394 11.2889 508 0.7877 0.4900 0.7877 0.8875
0.5394 11.3333 510 0.7623 0.4854 0.7623 0.8731

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run3_AugV5_k18_task2_organization

Finetuned
(4023)
this model