ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k14_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8135
  • Qwk: 0.6996
  • Mse: 0.8135
  • Rmse: 0.9020

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0426 2 2.2516 0.0334 2.2516 1.5005
No log 0.0851 4 1.4661 0.2248 1.4661 1.2108
No log 0.1277 6 1.5711 0.0927 1.5711 1.2534
No log 0.1702 8 1.6803 0.1997 1.6803 1.2963
No log 0.2128 10 1.9347 0.2232 1.9347 1.3910
No log 0.2553 12 1.9618 0.2381 1.9618 1.4007
No log 0.2979 14 1.7496 0.1785 1.7496 1.3227
No log 0.3404 16 1.7539 0.0669 1.7539 1.3243
No log 0.3830 18 1.8031 0.1296 1.8031 1.3428
No log 0.4255 20 2.1033 0.2262 2.1033 1.4503
No log 0.4681 22 2.2504 0.1799 2.2504 1.5001
No log 0.5106 24 2.0105 0.2290 2.0105 1.4179
No log 0.5532 26 1.8897 0.2681 1.8897 1.3746
No log 0.5957 28 1.8228 0.3089 1.8228 1.3501
No log 0.6383 30 1.7408 0.3374 1.7408 1.3194
No log 0.6809 32 1.7569 0.3393 1.7569 1.3255
No log 0.7234 34 1.7485 0.3643 1.7485 1.3223
No log 0.7660 36 1.7973 0.3481 1.7973 1.3407
No log 0.8085 38 2.0140 0.2537 2.0140 1.4192
No log 0.8511 40 2.1535 0.2364 2.1535 1.4675
No log 0.8936 42 2.3917 0.2206 2.3917 1.5465
No log 0.9362 44 2.0870 0.2314 2.0870 1.4446
No log 0.9787 46 1.7537 0.2874 1.7537 1.3243
No log 1.0213 48 1.7053 0.2706 1.7053 1.3059
No log 1.0638 50 1.9952 0.2662 1.9952 1.4125
No log 1.1064 52 2.9946 0.2169 2.9946 1.7305
No log 1.1489 54 3.5035 0.2121 3.5035 1.8718
No log 1.1915 56 3.2222 0.2375 3.2222 1.7950
No log 1.2340 58 2.5531 0.2312 2.5531 1.5978
No log 1.2766 60 2.3589 0.2345 2.3589 1.5359
No log 1.3191 62 1.8911 0.2550 1.8911 1.3752
No log 1.3617 64 1.6829 0.2694 1.6829 1.2973
No log 1.4043 66 1.7328 0.3295 1.7328 1.3164
No log 1.4468 68 2.1970 0.3171 2.1970 1.4822
No log 1.4894 70 2.5627 0.2757 2.5627 1.6008
No log 1.5319 72 2.7143 0.2543 2.7143 1.6475
No log 1.5745 74 2.6350 0.2437 2.6350 1.6233
No log 1.6170 76 2.4063 0.3037 2.4063 1.5512
No log 1.6596 78 2.1634 0.3698 2.1634 1.4708
No log 1.7021 80 1.8253 0.4147 1.8253 1.3510
No log 1.7447 82 1.8542 0.4158 1.8542 1.3617
No log 1.7872 84 2.1159 0.3911 2.1159 1.4546
No log 1.8298 86 2.3955 0.4130 2.3955 1.5477
No log 1.8723 88 2.2791 0.4054 2.2791 1.5097
No log 1.9149 90 1.7257 0.4610 1.7257 1.3137
No log 1.9574 92 1.4341 0.4669 1.4341 1.1975
No log 2.0 94 1.2612 0.4582 1.2612 1.1230
No log 2.0426 96 1.3339 0.5165 1.3339 1.1550
No log 2.0851 98 1.6194 0.4666 1.6194 1.2725
No log 2.1277 100 2.1400 0.4681 2.1400 1.4629
No log 2.1702 102 2.2807 0.4403 2.2807 1.5102
No log 2.2128 104 2.1156 0.4679 2.1156 1.4545
No log 2.2553 106 1.6329 0.4918 1.6329 1.2778
No log 2.2979 108 1.2786 0.4987 1.2786 1.1307
No log 2.3404 110 1.0265 0.5564 1.0265 1.0132
No log 2.3830 112 1.0613 0.5563 1.0613 1.0302
No log 2.4255 114 1.3373 0.4945 1.3373 1.1564
No log 2.4681 116 1.5728 0.4624 1.5728 1.2541
No log 2.5106 118 1.4193 0.4886 1.4193 1.1913
No log 2.5532 120 1.1127 0.5624 1.1127 1.0549
No log 2.5957 122 0.8834 0.6210 0.8834 0.9399
No log 2.6383 124 0.8469 0.6505 0.8469 0.9203
No log 2.6809 126 1.0007 0.6215 1.0007 1.0004
No log 2.7234 128 1.5348 0.5295 1.5348 1.2389
No log 2.7660 130 1.9796 0.4880 1.9796 1.4070
No log 2.8085 132 2.0292 0.4671 2.0292 1.4245
No log 2.8511 134 1.7881 0.4887 1.7881 1.3372
No log 2.8936 136 1.3948 0.5148 1.3948 1.1810
No log 2.9362 138 1.2059 0.5194 1.2059 1.0982
No log 2.9787 140 1.2195 0.5149 1.2195 1.1043
No log 3.0213 142 1.3911 0.5127 1.3911 1.1795
No log 3.0638 144 1.6336 0.4910 1.6336 1.2781
No log 3.1064 146 1.6457 0.5087 1.6457 1.2829
No log 3.1489 148 1.4100 0.5158 1.4100 1.1874
No log 3.1915 150 1.0668 0.5777 1.0668 1.0329
No log 3.2340 152 0.9018 0.6826 0.9018 0.9497
No log 3.2766 154 0.9229 0.6680 0.9229 0.9607
No log 3.3191 156 1.0632 0.6259 1.0632 1.0311
No log 3.3617 158 1.2269 0.5766 1.2269 1.1077
No log 3.4043 160 1.3373 0.5801 1.3373 1.1564
No log 3.4468 162 1.1661 0.6042 1.1661 1.0798
No log 3.4894 164 1.0108 0.6498 1.0108 1.0054
No log 3.5319 166 0.8856 0.6784 0.8856 0.9411
No log 3.5745 168 0.8978 0.6873 0.8978 0.9475
No log 3.6170 170 1.0488 0.5956 1.0488 1.0241
No log 3.6596 172 1.2351 0.5540 1.2351 1.1114
No log 3.7021 174 1.2451 0.5416 1.2451 1.1158
No log 3.7447 176 1.1136 0.5801 1.1136 1.0553
No log 3.7872 178 0.9498 0.6147 0.9498 0.9746
No log 3.8298 180 0.9402 0.6308 0.9402 0.9697
No log 3.8723 182 1.0822 0.5812 1.0822 1.0403
No log 3.9149 184 1.1305 0.5705 1.1305 1.0633
No log 3.9574 186 1.0303 0.6122 1.0303 1.0150
No log 4.0 188 1.0036 0.6280 1.0036 1.0018
No log 4.0426 190 0.9784 0.6176 0.9784 0.9891
No log 4.0851 192 1.0204 0.6140 1.0204 1.0102
No log 4.1277 194 1.0117 0.6248 1.0117 1.0058
No log 4.1702 196 1.0107 0.6403 1.0107 1.0053
No log 4.2128 198 0.9776 0.6517 0.9776 0.9887
No log 4.2553 200 1.0413 0.6340 1.0413 1.0205
No log 4.2979 202 1.0545 0.6074 1.0545 1.0269
No log 4.3404 204 1.1138 0.5979 1.1138 1.0553
No log 4.3830 206 1.0892 0.5979 1.0892 1.0436
No log 4.4255 208 0.9614 0.6616 0.9614 0.9805
No log 4.4681 210 0.8902 0.6672 0.8902 0.9435
No log 4.5106 212 0.8446 0.6910 0.8446 0.9190
No log 4.5532 214 0.7918 0.7051 0.7918 0.8898
No log 4.5957 216 0.8044 0.7051 0.8044 0.8969
No log 4.6383 218 0.9217 0.6916 0.9217 0.9601
No log 4.6809 220 1.0640 0.6356 1.0640 1.0315
No log 4.7234 222 1.0084 0.6302 1.0084 1.0042
No log 4.7660 224 0.8410 0.6942 0.8410 0.9171
No log 4.8085 226 0.7480 0.7044 0.7480 0.8649
No log 4.8511 228 0.7280 0.7044 0.7280 0.8532
No log 4.8936 230 0.7717 0.7083 0.7717 0.8785
No log 4.9362 232 0.8272 0.6900 0.8272 0.9095
No log 4.9787 234 0.9394 0.6622 0.9394 0.9692
No log 5.0213 236 0.9482 0.6581 0.9482 0.9737
No log 5.0638 238 0.8742 0.6714 0.8742 0.9350
No log 5.1064 240 0.8449 0.6773 0.8449 0.9192
No log 5.1489 242 0.8063 0.6864 0.8063 0.8979
No log 5.1915 244 0.7583 0.6994 0.7583 0.8708
No log 5.2340 246 0.7953 0.7030 0.7953 0.8918
No log 5.2766 248 0.8549 0.6874 0.8549 0.9246
No log 5.3191 250 0.8778 0.6714 0.8778 0.9369
No log 5.3617 252 0.9100 0.6706 0.9100 0.9539
No log 5.4043 254 0.8555 0.6768 0.8555 0.9249
No log 5.4468 256 0.7543 0.7120 0.7543 0.8685
No log 5.4894 258 0.7239 0.7035 0.7239 0.8508
No log 5.5319 260 0.7562 0.6970 0.7562 0.8696
No log 5.5745 262 0.8404 0.6590 0.8404 0.9167
No log 5.6170 264 0.9113 0.6655 0.9113 0.9546
No log 5.6596 266 0.8870 0.6646 0.8870 0.9418
No log 5.7021 268 0.8038 0.6894 0.8038 0.8966
No log 5.7447 270 0.7200 0.7208 0.7200 0.8485
No log 5.7872 272 0.7309 0.7109 0.7309 0.8549
No log 5.8298 274 0.8072 0.7015 0.8072 0.8984
No log 5.8723 276 0.9431 0.6673 0.9431 0.9712
No log 5.9149 278 1.0751 0.6604 1.0751 1.0368
No log 5.9574 280 1.0554 0.6500 1.0554 1.0273
No log 6.0 282 0.9427 0.6731 0.9427 0.9709
No log 6.0426 284 0.8644 0.6801 0.8644 0.9297
No log 6.0851 286 0.8322 0.6805 0.8322 0.9122
No log 6.1277 288 0.8256 0.6849 0.8256 0.9086
No log 6.1702 290 0.8490 0.6768 0.8490 0.9214
No log 6.2128 292 0.8983 0.6758 0.8983 0.9478
No log 6.2553 294 0.9866 0.6489 0.9866 0.9933
No log 6.2979 296 1.0027 0.6539 1.0027 1.0013
No log 6.3404 298 0.9158 0.6777 0.9158 0.9570
No log 6.3830 300 0.8220 0.6761 0.8220 0.9066
No log 6.4255 302 0.7361 0.7117 0.7361 0.8580
No log 6.4681 304 0.7153 0.7170 0.7153 0.8458
No log 6.5106 306 0.7696 0.7044 0.7696 0.8773
No log 6.5532 308 0.8331 0.6953 0.8331 0.9127
No log 6.5957 310 0.8131 0.7078 0.8131 0.9017
No log 6.6383 312 0.7359 0.7154 0.7359 0.8578
No log 6.6809 314 0.7014 0.7281 0.7014 0.8375
No log 6.7234 316 0.7295 0.7154 0.7295 0.8541
No log 6.7660 318 0.7264 0.7197 0.7264 0.8523
No log 6.8085 320 0.7272 0.7197 0.7272 0.8528
No log 6.8511 322 0.7555 0.7079 0.7555 0.8692
No log 6.8936 324 0.7590 0.7079 0.7590 0.8712
No log 6.9362 326 0.7698 0.6960 0.7698 0.8774
No log 6.9787 328 0.8175 0.7056 0.8175 0.9041
No log 7.0213 330 0.8343 0.7041 0.8343 0.9134
No log 7.0638 332 0.8439 0.7041 0.8439 0.9187
No log 7.1064 334 0.8512 0.6996 0.8512 0.9226
No log 7.1489 336 0.8451 0.6797 0.8451 0.9193
No log 7.1915 338 0.8304 0.6868 0.8304 0.9112
No log 7.2340 340 0.8045 0.6894 0.8045 0.8969
No log 7.2766 342 0.7974 0.6894 0.7974 0.8930
No log 7.3191 344 0.7712 0.7063 0.7712 0.8782
No log 7.3617 346 0.7605 0.7063 0.7605 0.8721
No log 7.4043 348 0.7608 0.7063 0.7608 0.8722
No log 7.4468 350 0.7508 0.7063 0.7508 0.8665
No log 7.4894 352 0.7370 0.7106 0.7370 0.8585
No log 7.5319 354 0.7388 0.7030 0.7388 0.8595
No log 7.5745 356 0.7640 0.6960 0.7640 0.8740
No log 7.6170 358 0.8393 0.7037 0.8393 0.9161
No log 7.6596 360 0.9418 0.6632 0.9418 0.9705
No log 7.7021 362 1.0225 0.6696 1.0225 1.0112
No log 7.7447 364 1.0235 0.6696 1.0235 1.0117
No log 7.7872 366 1.0046 0.6696 1.0046 1.0023
No log 7.8298 368 0.9449 0.6639 0.9449 0.9721
No log 7.8723 370 0.8582 0.6921 0.8582 0.9264
No log 7.9149 372 0.7891 0.7122 0.7891 0.8883
No log 7.9574 374 0.7721 0.7122 0.7721 0.8787
No log 8.0 376 0.7581 0.7122 0.7581 0.8707
No log 8.0426 378 0.7759 0.7122 0.7759 0.8809
No log 8.0851 380 0.7913 0.6999 0.7913 0.8896
No log 8.1277 382 0.8227 0.6758 0.8227 0.9070
No log 8.1702 384 0.8326 0.6731 0.8326 0.9125
No log 8.2128 386 0.8471 0.6731 0.8471 0.9204
No log 8.2553 388 0.8612 0.6773 0.8612 0.9280
No log 8.2979 390 0.8507 0.6861 0.8507 0.9224
No log 8.3404 392 0.8216 0.6844 0.8216 0.9064
No log 8.3830 394 0.7873 0.6967 0.7873 0.8873
No log 8.4255 396 0.7518 0.7106 0.7518 0.8671
No log 8.4681 398 0.7327 0.7106 0.7327 0.8560
No log 8.5106 400 0.7370 0.7106 0.7370 0.8585
No log 8.5532 402 0.7479 0.7106 0.7479 0.8648
No log 8.5957 404 0.7705 0.7166 0.7705 0.8778
No log 8.6383 406 0.8114 0.6892 0.8114 0.9008
No log 8.6809 408 0.8366 0.6932 0.8366 0.9147
No log 8.7234 410 0.8533 0.6932 0.8533 0.9238
No log 8.7660 412 0.8574 0.6932 0.8574 0.9259
No log 8.8085 414 0.8420 0.6932 0.8420 0.9176
No log 8.8511 416 0.8119 0.7034 0.8119 0.9011
No log 8.8936 418 0.7745 0.7018 0.7745 0.8800
No log 8.9362 420 0.7582 0.7109 0.7582 0.8707
No log 8.9787 422 0.7351 0.7106 0.7351 0.8574
No log 9.0213 424 0.7261 0.7070 0.7261 0.8521
No log 9.0638 426 0.7308 0.7070 0.7308 0.8549
No log 9.1064 428 0.7335 0.7070 0.7335 0.8564
No log 9.1489 430 0.7472 0.7106 0.7472 0.8644
No log 9.1915 432 0.7676 0.7114 0.7676 0.8761
No log 9.2340 434 0.7744 0.7114 0.7744 0.8800
No log 9.2766 436 0.7722 0.7190 0.7722 0.8788
No log 9.3191 438 0.7676 0.7106 0.7676 0.8761
No log 9.3617 440 0.7689 0.7109 0.7689 0.8769
No log 9.4043 442 0.7756 0.7109 0.7756 0.8807
No log 9.4468 444 0.7911 0.7041 0.7911 0.8894
No log 9.4894 446 0.8028 0.7132 0.8028 0.8960
No log 9.5319 448 0.8115 0.7112 0.8115 0.9009
No log 9.5745 450 0.8174 0.6996 0.8174 0.9041
No log 9.6170 452 0.8221 0.6996 0.8221 0.9067
No log 9.6596 454 0.8226 0.6996 0.8226 0.9070
No log 9.7021 456 0.8221 0.6996 0.8221 0.9067
No log 9.7447 458 0.8200 0.6996 0.8200 0.9055
No log 9.7872 460 0.8191 0.6996 0.8191 0.9050
No log 9.8298 462 0.8188 0.6996 0.8188 0.9048
No log 9.8723 464 0.8176 0.6996 0.8176 0.9042
No log 9.9149 466 0.8154 0.6996 0.8154 0.9030
No log 9.9574 468 0.8144 0.6996 0.8144 0.9024
No log 10.0 470 0.8135 0.6996 0.8135 0.9020

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k14_task5_organization

Finetuned
(4023)
this model