ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k14_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8136
  • Qwk: 0.6996
  • Mse: 0.8136
  • Rmse: 0.9020

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0426 2 2.2516 0.0334 2.2516 1.5005
No log 0.0851 4 1.4661 0.2248 1.4661 1.2108
No log 0.1277 6 1.5711 0.0927 1.5711 1.2534
No log 0.1702 8 1.6803 0.1997 1.6803 1.2963
No log 0.2128 10 1.9347 0.2232 1.9347 1.3910
No log 0.2553 12 1.9618 0.2381 1.9618 1.4007
No log 0.2979 14 1.7496 0.1785 1.7496 1.3227
No log 0.3404 16 1.7539 0.0669 1.7539 1.3243
No log 0.3830 18 1.8031 0.1296 1.8031 1.3428
No log 0.4255 20 2.1033 0.2262 2.1033 1.4503
No log 0.4681 22 2.2504 0.1799 2.2504 1.5001
No log 0.5106 24 2.0105 0.2290 2.0105 1.4179
No log 0.5532 26 1.8897 0.2681 1.8897 1.3746
No log 0.5957 28 1.8228 0.3089 1.8228 1.3501
No log 0.6383 30 1.7408 0.3374 1.7408 1.3194
No log 0.6809 32 1.7569 0.3393 1.7569 1.3255
No log 0.7234 34 1.7485 0.3643 1.7485 1.3223
No log 0.7660 36 1.7974 0.3481 1.7974 1.3407
No log 0.8085 38 2.0141 0.2537 2.0141 1.4192
No log 0.8511 40 2.1535 0.2364 2.1535 1.4675
No log 0.8936 42 2.3917 0.2206 2.3917 1.5465
No log 0.9362 44 2.0870 0.2314 2.0870 1.4446
No log 0.9787 46 1.7537 0.2874 1.7537 1.3243
No log 1.0213 48 1.7053 0.2706 1.7053 1.3059
No log 1.0638 50 1.9952 0.2662 1.9952 1.4125
No log 1.1064 52 2.9946 0.2169 2.9946 1.7305
No log 1.1489 54 3.5035 0.2121 3.5035 1.8718
No log 1.1915 56 3.2222 0.2375 3.2222 1.7950
No log 1.2340 58 2.5531 0.2312 2.5531 1.5978
No log 1.2766 60 2.3589 0.2345 2.3589 1.5359
No log 1.3191 62 1.8911 0.2550 1.8911 1.3752
No log 1.3617 64 1.6829 0.2694 1.6829 1.2973
No log 1.4043 66 1.7328 0.3295 1.7328 1.3164
No log 1.4468 68 2.1970 0.3171 2.1970 1.4822
No log 1.4894 70 2.5627 0.2757 2.5627 1.6008
No log 1.5319 72 2.7143 0.2543 2.7143 1.6475
No log 1.5745 74 2.6350 0.2437 2.6350 1.6233
No log 1.6170 76 2.4063 0.3037 2.4063 1.5512
No log 1.6596 78 2.1634 0.3698 2.1634 1.4708
No log 1.7021 80 1.8253 0.4147 1.8253 1.3510
No log 1.7447 82 1.8543 0.4085 1.8543 1.3617
No log 1.7872 84 2.1160 0.3911 2.1160 1.4546
No log 1.8298 86 2.3955 0.4130 2.3955 1.5477
No log 1.8723 88 2.2792 0.4054 2.2792 1.5097
No log 1.9149 90 1.7258 0.4610 1.7258 1.3137
No log 1.9574 92 1.4342 0.4669 1.4342 1.1976
No log 2.0 94 1.2613 0.4582 1.2613 1.1231
No log 2.0426 96 1.3340 0.5165 1.3340 1.1550
No log 2.0851 98 1.6194 0.4666 1.6194 1.2726
No log 2.1277 100 2.1400 0.4681 2.1400 1.4629
No log 2.1702 102 2.2808 0.4403 2.2808 1.5102
No log 2.2128 104 2.1157 0.4679 2.1157 1.4546
No log 2.2553 106 1.6330 0.4918 1.6330 1.2779
No log 2.2979 108 1.2787 0.4987 1.2787 1.1308
No log 2.3404 110 1.0265 0.5564 1.0265 1.0132
No log 2.3830 112 1.0613 0.5563 1.0613 1.0302
No log 2.4255 114 1.3373 0.4945 1.3373 1.1564
No log 2.4681 116 1.5728 0.4624 1.5728 1.2541
No log 2.5106 118 1.4193 0.4886 1.4193 1.1913
No log 2.5532 120 1.1128 0.5624 1.1128 1.0549
No log 2.5957 122 0.8834 0.6210 0.8834 0.9399
No log 2.6383 124 0.8469 0.6505 0.8469 0.9203
No log 2.6809 126 1.0007 0.6215 1.0007 1.0004
No log 2.7234 128 1.5347 0.5295 1.5347 1.2388
No log 2.7660 130 1.9795 0.4880 1.9795 1.4070
No log 2.8085 132 2.0292 0.4671 2.0292 1.4245
No log 2.8511 134 1.7880 0.4887 1.7880 1.3372
No log 2.8936 136 1.3947 0.5148 1.3947 1.1810
No log 2.9362 138 1.2059 0.5194 1.2059 1.0981
No log 2.9787 140 1.2194 0.5149 1.2194 1.1043
No log 3.0213 142 1.3911 0.5127 1.3911 1.1794
No log 3.0638 144 1.6335 0.4910 1.6335 1.2781
No log 3.1064 146 1.6456 0.5087 1.6456 1.2828
No log 3.1489 148 1.4099 0.5158 1.4099 1.1874
No log 3.1915 150 1.0668 0.5777 1.0668 1.0329
No log 3.2340 152 0.9018 0.6826 0.9018 0.9496
No log 3.2766 154 0.9229 0.6680 0.9229 0.9607
No log 3.3191 156 1.0632 0.6259 1.0632 1.0311
No log 3.3617 158 1.2270 0.5766 1.2270 1.1077
No log 3.4043 160 1.3373 0.5801 1.3373 1.1564
No log 3.4468 162 1.1661 0.6042 1.1661 1.0799
No log 3.4894 164 1.0108 0.6498 1.0108 1.0054
No log 3.5319 166 0.8856 0.6784 0.8856 0.9411
No log 3.5745 168 0.8978 0.6873 0.8978 0.9475
No log 3.6170 170 1.0488 0.5956 1.0488 1.0241
No log 3.6596 172 1.2351 0.5540 1.2351 1.1114
No log 3.7021 174 1.2451 0.5416 1.2451 1.1159
No log 3.7447 176 1.1137 0.5801 1.1137 1.0553
No log 3.7872 178 0.9499 0.6147 0.9499 0.9746
No log 3.8298 180 0.9403 0.6308 0.9403 0.9697
No log 3.8723 182 1.0823 0.5812 1.0823 1.0403
No log 3.9149 184 1.1306 0.5705 1.1306 1.0633
No log 3.9574 186 1.0304 0.6122 1.0304 1.0151
No log 4.0 188 1.0037 0.6280 1.0037 1.0018
No log 4.0426 190 0.9784 0.6176 0.9784 0.9891
No log 4.0851 192 1.0204 0.6140 1.0204 1.0102
No log 4.1277 194 1.0117 0.6248 1.0117 1.0059
No log 4.1702 196 1.0108 0.6403 1.0108 1.0054
No log 4.2128 198 0.9777 0.6517 0.9777 0.9888
No log 4.2553 200 1.0415 0.6340 1.0415 1.0205
No log 4.2979 202 1.0545 0.6074 1.0545 1.0269
No log 4.3404 204 1.1137 0.5979 1.1137 1.0553
No log 4.3830 206 1.0890 0.5979 1.0890 1.0436
No log 4.4255 208 0.9612 0.6616 0.9612 0.9804
No log 4.4681 210 0.8901 0.6672 0.8901 0.9434
No log 4.5106 212 0.8446 0.6910 0.8446 0.9190
No log 4.5532 214 0.7917 0.7051 0.7917 0.8898
No log 4.5957 216 0.8044 0.7051 0.8044 0.8969
No log 4.6383 218 0.9218 0.6916 0.9218 0.9601
No log 4.6809 220 1.0639 0.6356 1.0639 1.0315
No log 4.7234 222 1.0083 0.6302 1.0083 1.0041
No log 4.7660 224 0.8409 0.6942 0.8409 0.9170
No log 4.8085 226 0.7479 0.7044 0.7479 0.8648
No log 4.8511 228 0.7280 0.7044 0.7280 0.8532
No log 4.8936 230 0.7717 0.7083 0.7717 0.8785
No log 4.9362 232 0.8272 0.6900 0.8272 0.9095
No log 4.9787 234 0.9395 0.6622 0.9395 0.9693
No log 5.0213 236 0.9482 0.6581 0.9482 0.9737
No log 5.0638 238 0.8741 0.6714 0.8741 0.9350
No log 5.1064 240 0.8448 0.6773 0.8448 0.9191
No log 5.1489 242 0.8063 0.6864 0.8063 0.8979
No log 5.1915 244 0.7584 0.6994 0.7584 0.8708
No log 5.2340 246 0.7954 0.7030 0.7954 0.8918
No log 5.2766 248 0.8551 0.6874 0.8551 0.9247
No log 5.3191 250 0.8776 0.6714 0.8776 0.9368
No log 5.3617 252 0.9097 0.6706 0.9097 0.9538
No log 5.4043 254 0.8552 0.6768 0.8552 0.9248
No log 5.4468 256 0.7542 0.7120 0.7542 0.8685
No log 5.4894 258 0.7240 0.7035 0.7240 0.8509
No log 5.5319 260 0.7565 0.6843 0.7565 0.8698
No log 5.5745 262 0.8408 0.6590 0.8408 0.9170
No log 5.6170 264 0.9115 0.6655 0.9115 0.9547
No log 5.6596 266 0.8868 0.6646 0.8868 0.9417
No log 5.7021 268 0.8035 0.6894 0.8035 0.8964
No log 5.7447 270 0.7197 0.7208 0.7197 0.8483
No log 5.7872 272 0.7307 0.7109 0.7307 0.8548
No log 5.8298 274 0.8073 0.7015 0.8073 0.8985
No log 5.8723 276 0.9435 0.6673 0.9435 0.9713
No log 5.9149 278 1.0754 0.6604 1.0754 1.0370
No log 5.9574 280 1.0557 0.6500 1.0557 1.0275
No log 6.0 282 0.9430 0.6731 0.9430 0.9711
No log 6.0426 284 0.8645 0.6801 0.8645 0.9298
No log 6.0851 286 0.8322 0.6805 0.8322 0.9122
No log 6.1277 288 0.8255 0.6849 0.8255 0.9086
No log 6.1702 290 0.8489 0.6768 0.8489 0.9213
No log 6.2128 292 0.8981 0.6758 0.8981 0.9477
No log 6.2553 294 0.9865 0.6489 0.9865 0.9933
No log 6.2979 296 1.0028 0.6539 1.0028 1.0014
No log 6.3404 298 0.9161 0.6777 0.9161 0.9571
No log 6.3830 300 0.8222 0.6761 0.8222 0.9067
No log 6.4255 302 0.7362 0.7117 0.7362 0.8580
No log 6.4681 304 0.7154 0.7170 0.7154 0.8458
No log 6.5106 306 0.7696 0.7044 0.7696 0.8773
No log 6.5532 308 0.8331 0.6953 0.8331 0.9127
No log 6.5957 310 0.8131 0.7078 0.8131 0.9017
No log 6.6383 312 0.7359 0.7154 0.7359 0.8579
No log 6.6809 314 0.7014 0.7281 0.7014 0.8375
No log 6.7234 316 0.7296 0.7154 0.7296 0.8542
No log 6.7660 318 0.7264 0.7197 0.7264 0.8523
No log 6.8085 320 0.7274 0.7197 0.7274 0.8529
No log 6.8511 322 0.7557 0.7079 0.7557 0.8693
No log 6.8936 324 0.7592 0.7079 0.7592 0.8713
No log 6.9362 326 0.7701 0.6960 0.7701 0.8776
No log 6.9787 328 0.8176 0.7056 0.8176 0.9042
No log 7.0213 330 0.8343 0.7041 0.8343 0.9134
No log 7.0638 332 0.8439 0.7041 0.8439 0.9186
No log 7.1064 334 0.8510 0.6996 0.8510 0.9225
No log 7.1489 336 0.8449 0.6797 0.8449 0.9192
No log 7.1915 338 0.8301 0.6868 0.8301 0.9111
No log 7.2340 340 0.8043 0.6894 0.8043 0.8968
No log 7.2766 342 0.7973 0.6894 0.7973 0.8929
No log 7.3191 344 0.7713 0.7063 0.7713 0.8782
No log 7.3617 346 0.7607 0.7063 0.7607 0.8722
No log 7.4043 348 0.7610 0.7063 0.7610 0.8724
No log 7.4468 350 0.7510 0.7063 0.7510 0.8666
No log 7.4894 352 0.7371 0.7106 0.7371 0.8585
No log 7.5319 354 0.7388 0.7030 0.7388 0.8596
No log 7.5745 356 0.7640 0.6960 0.7640 0.8740
No log 7.6170 358 0.8393 0.7037 0.8393 0.9161
No log 7.6596 360 0.9417 0.6632 0.9417 0.9704
No log 7.7021 362 1.0224 0.6696 1.0224 1.0112
No log 7.7447 364 1.0234 0.6696 1.0234 1.0116
No log 7.7872 366 1.0045 0.6696 1.0045 1.0023
No log 7.8298 368 0.9450 0.6639 0.9450 0.9721
No log 7.8723 370 0.8583 0.6921 0.8583 0.9265
No log 7.9149 372 0.7893 0.7122 0.7893 0.8884
No log 7.9574 374 0.7724 0.7122 0.7724 0.8788
No log 8.0 376 0.7584 0.7122 0.7584 0.8708
No log 8.0426 378 0.7762 0.7122 0.7762 0.8810
No log 8.0851 380 0.7914 0.6999 0.7914 0.8896
No log 8.1277 382 0.8226 0.6758 0.8226 0.9070
No log 8.1702 384 0.8325 0.6731 0.8325 0.9124
No log 8.2128 386 0.8469 0.6731 0.8469 0.9203
No log 8.2553 388 0.8610 0.6773 0.8610 0.9279
No log 8.2979 390 0.8505 0.6861 0.8505 0.9222
No log 8.3404 392 0.8214 0.6844 0.8214 0.9063
No log 8.3830 394 0.7871 0.6967 0.7871 0.8872
No log 8.4255 396 0.7516 0.7106 0.7516 0.8670
No log 8.4681 398 0.7325 0.7106 0.7325 0.8559
No log 8.5106 400 0.7369 0.7106 0.7369 0.8584
No log 8.5532 402 0.7478 0.7106 0.7478 0.8648
No log 8.5957 404 0.7705 0.7166 0.7705 0.8778
No log 8.6383 406 0.8113 0.6892 0.8113 0.9007
No log 8.6809 408 0.8366 0.6932 0.8366 0.9147
No log 8.7234 410 0.8534 0.6932 0.8534 0.9238
No log 8.7660 412 0.8575 0.6932 0.8575 0.9260
No log 8.8085 414 0.8422 0.6932 0.8422 0.9177
No log 8.8511 416 0.8121 0.7034 0.8121 0.9012
No log 8.8936 418 0.7746 0.7018 0.7746 0.8801
No log 8.9362 420 0.7584 0.7109 0.7584 0.8708
No log 8.9787 422 0.7353 0.7106 0.7353 0.8575
No log 9.0213 424 0.7263 0.7070 0.7263 0.8522
No log 9.0638 426 0.7310 0.7070 0.7310 0.8550
No log 9.1064 428 0.7336 0.7070 0.7336 0.8565
No log 9.1489 430 0.7473 0.7106 0.7473 0.8645
No log 9.1915 432 0.7677 0.7114 0.7677 0.8762
No log 9.2340 434 0.7745 0.7114 0.7745 0.8801
No log 9.2766 436 0.7724 0.7190 0.7724 0.8788
No log 9.3191 438 0.7678 0.7106 0.7678 0.8762
No log 9.3617 440 0.7690 0.7109 0.7690 0.8770
No log 9.4043 442 0.7757 0.7109 0.7757 0.8807
No log 9.4468 444 0.7911 0.7041 0.7911 0.8894
No log 9.4894 446 0.8028 0.7132 0.8028 0.8960
No log 9.5319 448 0.8116 0.7112 0.8116 0.9009
No log 9.5745 450 0.8174 0.6996 0.8174 0.9041
No log 9.6170 452 0.8222 0.6996 0.8222 0.9067
No log 9.6596 454 0.8227 0.6996 0.8227 0.9070
No log 9.7021 456 0.8222 0.6996 0.8222 0.9068
No log 9.7447 458 0.8201 0.6996 0.8201 0.9056
No log 9.7872 460 0.8192 0.6996 0.8192 0.9051
No log 9.8298 462 0.8188 0.6996 0.8188 0.9049
No log 9.8723 464 0.8177 0.6996 0.8177 0.9042
No log 9.9149 466 0.8155 0.6996 0.8155 0.9030
No log 9.9574 468 0.8145 0.6996 0.8145 0.9025
No log 10.0 470 0.8136 0.6996 0.8136 0.9020

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k14_task5_organization

Finetuned
(4023)
this model