ArabicNewSplits6_FineTuningAraBERTFreeze_run3_AugV5_k17_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7989
  • Qwk: 0.3997
  • Mse: 0.7989
  • Rmse: 0.8938

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 16
  • eval_batch_size: 16
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0476 2 6.4407 -0.0278 6.4407 2.5379
No log 0.0952 4 4.2959 -0.0283 4.2959 2.0727
No log 0.1429 6 2.9670 -0.0006 2.9670 1.7225
No log 0.1905 8 2.1222 0.0146 2.1222 1.4568
No log 0.2381 10 1.5838 -0.0355 1.5838 1.2585
No log 0.2857 12 1.5766 -0.0606 1.5766 1.2556
No log 0.3333 14 1.6883 -0.0794 1.6883 1.2993
No log 0.3810 16 1.8362 -0.0877 1.8362 1.3551
No log 0.4286 18 1.7557 -0.0889 1.7557 1.3250
No log 0.4762 20 1.3637 0.0125 1.3637 1.1678
No log 0.5238 22 0.9524 0.0197 0.9524 0.9759
No log 0.5714 24 0.7625 0.1611 0.7625 0.8732
No log 0.6190 26 0.7096 0.2713 0.7096 0.8424
No log 0.6667 28 0.7239 0.3077 0.7239 0.8508
No log 0.7143 30 0.7999 0.2366 0.7999 0.8944
No log 0.7619 32 0.8715 0.1312 0.8715 0.9335
No log 0.8095 34 0.9483 0.1596 0.9483 0.9738
No log 0.8571 36 0.9556 0.1596 0.9556 0.9776
No log 0.9048 38 0.9258 0.2026 0.9258 0.9622
No log 0.9524 40 0.9230 0.2464 0.9230 0.9607
No log 1.0 42 0.8507 0.2089 0.8507 0.9223
No log 1.0476 44 0.8611 0.2987 0.8611 0.9279
No log 1.0952 46 0.9101 0.2996 0.9101 0.9540
No log 1.1429 48 0.9211 0.2996 0.9211 0.9597
No log 1.1905 50 0.8458 0.3546 0.8458 0.9197
No log 1.2381 52 0.7964 0.3788 0.7964 0.8924
No log 1.2857 54 0.6991 0.4410 0.6991 0.8361
No log 1.3333 56 0.6309 0.4440 0.6309 0.7943
No log 1.3810 58 0.6072 0.4071 0.6072 0.7792
No log 1.4286 60 0.5968 0.4395 0.5968 0.7725
No log 1.4762 62 0.6054 0.4259 0.6054 0.7781
No log 1.5238 64 0.6328 0.4561 0.6328 0.7955
No log 1.5714 66 0.6640 0.4481 0.6640 0.8148
No log 1.6190 68 0.7704 0.4114 0.7704 0.8777
No log 1.6667 70 0.7929 0.4181 0.7929 0.8904
No log 1.7143 72 0.7662 0.4137 0.7662 0.8753
No log 1.7619 74 0.6936 0.4593 0.6936 0.8329
No log 1.8095 76 0.6280 0.4761 0.6280 0.7925
No log 1.8571 78 0.6388 0.4624 0.6388 0.7993
No log 1.9048 80 0.6511 0.4406 0.6511 0.8069
No log 1.9524 82 0.7075 0.471 0.7075 0.8411
No log 2.0 84 0.7919 0.4417 0.7919 0.8899
No log 2.0476 86 0.7968 0.4438 0.7968 0.8926
No log 2.0952 88 0.7144 0.4847 0.7144 0.8453
No log 2.1429 90 0.6310 0.4487 0.6310 0.7944
No log 2.1905 92 0.6004 0.4535 0.6004 0.7749
No log 2.2381 94 0.6141 0.4666 0.6141 0.7837
No log 2.2857 96 0.7008 0.4988 0.7008 0.8372
No log 2.3333 98 0.8422 0.4793 0.8422 0.9177
No log 2.3810 100 0.8156 0.4702 0.8156 0.9031
No log 2.4286 102 0.6892 0.4798 0.6892 0.8302
No log 2.4762 104 0.6250 0.4971 0.6250 0.7906
No log 2.5238 106 0.6203 0.4908 0.6203 0.7876
No log 2.5714 108 0.6502 0.5142 0.6502 0.8063
No log 2.6190 110 0.6725 0.5344 0.6725 0.8201
No log 2.6667 112 0.7104 0.5048 0.7104 0.8429
No log 2.7143 114 0.7137 0.4992 0.7137 0.8448
No log 2.7619 116 0.6627 0.5169 0.6627 0.8141
No log 2.8095 118 0.5996 0.4974 0.5996 0.7743
No log 2.8571 120 0.5934 0.5183 0.5934 0.7703
No log 2.9048 122 0.5789 0.5070 0.5789 0.7608
No log 2.9524 124 0.5985 0.4954 0.5985 0.7736
No log 3.0 126 0.6539 0.5290 0.6539 0.8087
No log 3.0476 128 0.6934 0.5211 0.6934 0.8327
No log 3.0952 130 0.6944 0.5325 0.6944 0.8333
No log 3.1429 132 0.7323 0.5325 0.7323 0.8558
No log 3.1905 134 0.7399 0.4981 0.7399 0.8602
No log 3.2381 136 0.6915 0.5353 0.6915 0.8315
No log 3.2857 138 0.6703 0.5116 0.6703 0.8187
No log 3.3333 140 0.6744 0.5126 0.6744 0.8212
No log 3.3810 142 0.7049 0.4810 0.7049 0.8396
No log 3.4286 144 0.7365 0.5018 0.7365 0.8582
No log 3.4762 146 0.7089 0.5011 0.7089 0.8420
No log 3.5238 148 0.6599 0.5006 0.6599 0.8123
No log 3.5714 150 0.6480 0.5111 0.6480 0.8050
No log 3.6190 152 0.6474 0.5055 0.6474 0.8046
No log 3.6667 154 0.6780 0.4967 0.6780 0.8234
No log 3.7143 156 0.7136 0.4860 0.7136 0.8447
No log 3.7619 158 0.6925 0.5044 0.6925 0.8322
No log 3.8095 160 0.6781 0.5067 0.6781 0.8235
No log 3.8571 162 0.6695 0.5020 0.6695 0.8182
No log 3.9048 164 0.6853 0.5011 0.6853 0.8279
No log 3.9524 166 0.7268 0.5086 0.7268 0.8525
No log 4.0 168 0.8403 0.5005 0.8403 0.9167
No log 4.0476 170 0.8665 0.4904 0.8665 0.9308
No log 4.0952 172 0.7790 0.4822 0.7790 0.8826
No log 4.1429 174 0.7326 0.4593 0.7326 0.8559
No log 4.1905 176 0.7304 0.4598 0.7304 0.8546
No log 4.2381 178 0.7633 0.4756 0.7633 0.8737
No log 4.2857 180 0.8340 0.4975 0.8340 0.9132
No log 4.3333 182 0.8386 0.4975 0.8386 0.9158
No log 4.3810 184 0.8118 0.4928 0.8118 0.9010
No log 4.4286 186 0.7661 0.5021 0.7661 0.8752
No log 4.4762 188 0.7632 0.4714 0.7632 0.8736
No log 4.5238 190 0.7893 0.4488 0.7893 0.8885
No log 4.5714 192 0.7677 0.4547 0.7677 0.8762
No log 4.6190 194 0.7388 0.4521 0.7388 0.8595
No log 4.6667 196 0.8036 0.5235 0.8036 0.8964
No log 4.7143 198 0.9775 0.4485 0.9775 0.9887
No log 4.7619 200 1.0399 0.4208 1.0399 1.0198
No log 4.8095 202 0.9004 0.4705 0.9004 0.9489
No log 4.8571 204 0.7254 0.5320 0.7254 0.8517
No log 4.9048 206 0.6980 0.4752 0.6980 0.8355
No log 4.9524 208 0.7034 0.4835 0.7034 0.8387
No log 5.0 210 0.6931 0.4752 0.6931 0.8325
No log 5.0476 212 0.6948 0.5109 0.6948 0.8335
No log 5.0952 214 0.7482 0.5214 0.7482 0.8650
No log 5.1429 216 0.8520 0.4770 0.8520 0.9230
No log 5.1905 218 0.8417 0.4689 0.8417 0.9174
No log 5.2381 220 0.8049 0.5026 0.8049 0.8971
No log 5.2857 222 0.7779 0.4657 0.7779 0.8820
No log 5.3333 224 0.7904 0.4569 0.7904 0.8891
No log 5.3810 226 0.8283 0.4529 0.8283 0.9101
No log 5.4286 228 0.8744 0.4604 0.8744 0.9351
No log 5.4762 230 0.8865 0.4812 0.8865 0.9415
No log 5.5238 232 0.8649 0.4934 0.8649 0.9300
No log 5.5714 234 0.8531 0.5082 0.8531 0.9237
No log 5.6190 236 0.8051 0.4926 0.8051 0.8973
No log 5.6667 238 0.7781 0.4778 0.7781 0.8821
No log 5.7143 240 0.8001 0.4560 0.8001 0.8945
No log 5.7619 242 0.8026 0.4379 0.8026 0.8959
No log 5.8095 244 0.7429 0.4688 0.7429 0.8619
No log 5.8571 246 0.6759 0.5090 0.6759 0.8221
No log 5.9048 248 0.7157 0.4703 0.7157 0.8460
No log 5.9524 250 0.7357 0.4737 0.7357 0.8577
No log 6.0 252 0.6974 0.5132 0.6974 0.8351
No log 6.0476 254 0.6989 0.4937 0.6989 0.8360
No log 6.0952 256 0.7011 0.4760 0.7011 0.8373
No log 6.1429 258 0.7215 0.4606 0.7215 0.8494
No log 6.1905 260 0.7974 0.5039 0.7974 0.8930
No log 6.2381 262 0.9173 0.4729 0.9173 0.9578
No log 6.2857 264 0.9814 0.4552 0.9814 0.9906
No log 6.3333 266 0.9126 0.4947 0.9126 0.9553
No log 6.3810 268 0.8478 0.4947 0.8478 0.9208
No log 6.4286 270 0.8519 0.4507 0.8519 0.9230
No log 6.4762 272 0.8500 0.4619 0.8500 0.9220
No log 6.5238 274 0.8422 0.4649 0.8422 0.9177
No log 6.5714 276 0.8390 0.4886 0.8390 0.9160
No log 6.6190 278 0.8102 0.4726 0.8102 0.9001
No log 6.6667 280 0.7874 0.4791 0.7874 0.8874
No log 6.7143 282 0.7953 0.5037 0.7953 0.8918
No log 6.7619 284 0.8048 0.5044 0.8048 0.8971
No log 6.8095 286 0.8052 0.5087 0.8052 0.8974
No log 6.8571 288 0.7919 0.4953 0.7919 0.8899
No log 6.9048 290 0.7968 0.4836 0.7968 0.8926
No log 6.9524 292 0.8176 0.4821 0.8176 0.9042
No log 7.0 294 0.8255 0.4837 0.8255 0.9086
No log 7.0476 296 0.8130 0.4850 0.8130 0.9017
No log 7.0952 298 0.8048 0.4961 0.8048 0.8971
No log 7.1429 300 0.7853 0.5037 0.7853 0.8862
No log 7.1905 302 0.7663 0.4822 0.7663 0.8754
No log 7.2381 304 0.7654 0.4593 0.7654 0.8749
No log 7.2857 306 0.8274 0.4924 0.8274 0.9096
No log 7.3333 308 0.8388 0.4819 0.8388 0.9159
No log 7.3810 310 0.7734 0.4480 0.7734 0.8794
No log 7.4286 312 0.7261 0.4539 0.7261 0.8521
No log 7.4762 314 0.7187 0.4669 0.7187 0.8478
No log 7.5238 316 0.7134 0.4710 0.7134 0.8446
No log 7.5714 318 0.7153 0.4524 0.7153 0.8458
No log 7.6190 320 0.7440 0.4582 0.7440 0.8625
No log 7.6667 322 0.7744 0.4711 0.7744 0.8800
No log 7.7143 324 0.7780 0.4715 0.7780 0.8820
No log 7.7619 326 0.7768 0.4952 0.7768 0.8814
No log 7.8095 328 0.7836 0.5196 0.7836 0.8852
No log 7.8571 330 0.7931 0.5057 0.7931 0.8906
No log 7.9048 332 0.8104 0.4869 0.8104 0.9002
No log 7.9524 334 0.8887 0.4778 0.8887 0.9427
No log 8.0 336 0.8948 0.4480 0.8948 0.9460
No log 8.0476 338 0.8191 0.4805 0.8191 0.9050
No log 8.0952 340 0.7669 0.5236 0.7669 0.8757
No log 8.1429 342 0.7960 0.4827 0.7960 0.8922
No log 8.1905 344 0.7939 0.4938 0.7939 0.8910
No log 8.2381 346 0.7689 0.5063 0.7689 0.8769
No log 8.2857 348 0.7435 0.5033 0.7435 0.8623
No log 8.3333 350 0.7893 0.4868 0.7893 0.8884
No log 8.3810 352 0.8299 0.4309 0.8299 0.9110
No log 8.4286 354 0.7878 0.4441 0.7878 0.8876
No log 8.4762 356 0.7479 0.5120 0.7479 0.8648
No log 8.5238 358 0.7576 0.5021 0.7576 0.8704
No log 8.5714 360 0.7645 0.5013 0.7645 0.8744
No log 8.6190 362 0.7611 0.5149 0.7611 0.8724
No log 8.6667 364 0.7816 0.4765 0.7816 0.8841
No log 8.7143 366 0.8430 0.4702 0.8430 0.9182
No log 8.7619 368 0.8386 0.4686 0.8386 0.9157
No log 8.8095 370 0.7883 0.4524 0.7883 0.8878
No log 8.8571 372 0.7739 0.4734 0.7739 0.8797
No log 8.9048 374 0.7811 0.4759 0.7811 0.8838
No log 8.9524 376 0.8065 0.4765 0.8065 0.8981
No log 9.0 378 0.9069 0.4435 0.9069 0.9523
No log 9.0476 380 0.9747 0.4542 0.9747 0.9873
No log 9.0952 382 0.9221 0.4417 0.9221 0.9603
No log 9.1429 384 0.8306 0.4566 0.8306 0.9114
No log 9.1905 386 0.8083 0.4560 0.8083 0.8990
No log 9.2381 388 0.8193 0.4707 0.8193 0.9051
No log 9.2857 390 0.8289 0.4599 0.8289 0.9104
No log 9.3333 392 0.8179 0.4651 0.8179 0.9044
No log 9.3810 394 0.8182 0.4535 0.8182 0.9046
No log 9.4286 396 0.8523 0.4253 0.8523 0.9232
No log 9.4762 398 0.8313 0.4048 0.8313 0.9118
No log 9.5238 400 0.7779 0.4609 0.7779 0.8820
No log 9.5714 402 0.7500 0.5352 0.7500 0.8660
No log 9.6190 404 0.7396 0.5352 0.7396 0.8600
No log 9.6667 406 0.7317 0.5182 0.7317 0.8554
No log 9.7143 408 0.7365 0.5248 0.7365 0.8582
No log 9.7619 410 0.7370 0.5017 0.7370 0.8585
No log 9.8095 412 0.7444 0.4414 0.7444 0.8628
No log 9.8571 414 0.7319 0.4679 0.7319 0.8555
No log 9.9048 416 0.7190 0.5104 0.7190 0.8479
No log 9.9524 418 0.7219 0.5195 0.7219 0.8496
No log 10.0 420 0.7283 0.4992 0.7283 0.8534
No log 10.0476 422 0.7664 0.4302 0.7664 0.8755
No log 10.0952 424 0.8801 0.4441 0.8801 0.9382
No log 10.1429 426 0.9596 0.4279 0.9596 0.9796
No log 10.1905 428 0.9117 0.4456 0.9117 0.9548
No log 10.2381 430 0.8286 0.3979 0.8286 0.9103
No log 10.2857 432 0.7931 0.4628 0.7931 0.8906
No log 10.3333 434 0.7954 0.4624 0.7954 0.8918
No log 10.3810 436 0.8310 0.4233 0.8310 0.9116
No log 10.4286 438 0.8460 0.4013 0.8460 0.9198
No log 10.4762 440 0.8229 0.4382 0.8229 0.9071
No log 10.5238 442 0.7940 0.4918 0.7940 0.8910
No log 10.5714 444 0.7954 0.5118 0.7954 0.8919
No log 10.6190 446 0.8117 0.5 0.8117 0.9010
No log 10.6667 448 0.8017 0.5206 0.8017 0.8954
No log 10.7143 450 0.7784 0.4900 0.7784 0.8823
No log 10.7619 452 0.8190 0.4046 0.8190 0.9050
No log 10.8095 454 0.8930 0.4070 0.8930 0.9450
No log 10.8571 456 0.8781 0.4068 0.8781 0.9371
No log 10.9048 458 0.7985 0.4142 0.7985 0.8936
No log 10.9524 460 0.7508 0.4799 0.7508 0.8665
No log 11.0 462 0.7809 0.5050 0.7809 0.8837
No log 11.0476 464 0.8036 0.5211 0.8036 0.8964
No log 11.0952 466 0.7888 0.5035 0.7888 0.8882
No log 11.1429 468 0.7776 0.4458 0.7776 0.8818
No log 11.1905 470 0.8632 0.4241 0.8632 0.9291
No log 11.2381 472 1.1169 0.3983 1.1169 1.0568
No log 11.2857 474 1.2259 0.3776 1.2259 1.1072
No log 11.3333 476 1.1258 0.3939 1.1258 1.0611
No log 11.3810 478 0.9260 0.4486 0.9260 0.9623
No log 11.4286 480 0.7787 0.4562 0.7787 0.8824
No log 11.4762 482 0.7524 0.4860 0.7524 0.8674
No log 11.5238 484 0.7580 0.4860 0.7580 0.8706
No log 11.5714 486 0.7674 0.4770 0.7674 0.8760
No log 11.6190 488 0.7829 0.4686 0.7829 0.8848
No log 11.6667 490 0.7956 0.4628 0.7956 0.8920
No log 11.7143 492 0.8037 0.4655 0.8037 0.8965
No log 11.7619 494 0.8150 0.4811 0.8150 0.9028
No log 11.8095 496 0.8314 0.4489 0.8314 0.9118
No log 11.8571 498 0.8320 0.4489 0.8320 0.9121
0.5522 11.9048 500 0.8317 0.4469 0.8317 0.9120
0.5522 11.9524 502 0.8484 0.4100 0.8484 0.9211
0.5522 12.0 504 0.8428 0.4338 0.8428 0.9181
0.5522 12.0476 506 0.8038 0.4354 0.8038 0.8965
0.5522 12.0952 508 0.7697 0.4497 0.7697 0.8773
0.5522 12.1429 510 0.7592 0.4627 0.7592 0.8713
0.5522 12.1905 512 0.7623 0.4500 0.7623 0.8731
0.5522 12.2381 514 0.7594 0.4373 0.7594 0.8714
0.5522 12.2857 516 0.7560 0.4427 0.7560 0.8695
0.5522 12.3333 518 0.7544 0.4809 0.7544 0.8686
0.5522 12.3810 520 0.7671 0.4809 0.7671 0.8758
0.5522 12.4286 522 0.7780 0.4456 0.7780 0.8820
0.5522 12.4762 524 0.7817 0.4456 0.7817 0.8842
0.5522 12.5238 526 0.7900 0.4281 0.7900 0.8888
0.5522 12.5714 528 0.8157 0.4212 0.8157 0.9031
0.5522 12.6190 530 0.7989 0.3997 0.7989 0.8938

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.1+cu121
  • Datasets 3.2.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERTFreeze_run3_AugV5_k17_task2_organization

Finetuned
(4023)
this model