ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k7_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5342
  • Qwk: 0.4105
  • Mse: 0.5342
  • Rmse: 0.7309

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 3.1992 -0.0041 3.1992 1.7886
No log 0.1111 4 1.5636 0.0255 1.5636 1.2504
No log 0.1667 6 1.1570 0.0075 1.1570 1.0756
No log 0.2222 8 0.6881 0.0476 0.6881 0.8295
No log 0.2778 10 1.0431 0.1169 1.0431 1.0213
No log 0.3333 12 0.8218 0.1644 0.8218 0.9066
No log 0.3889 14 0.7110 0.0769 0.7110 0.8432
No log 0.4444 16 0.8654 0.1515 0.8654 0.9302
No log 0.5 18 0.6017 0.1407 0.6017 0.7757
No log 0.5556 20 1.0585 0.0522 1.0585 1.0288
No log 0.6111 22 1.4266 0.0 1.4266 1.1944
No log 0.6667 24 0.9701 0.0345 0.9701 0.9849
No log 0.7222 26 0.6078 -0.0159 0.6078 0.7796
No log 0.7778 28 0.6106 0.0 0.6106 0.7814
No log 0.8333 30 0.7426 0.0815 0.7426 0.8618
No log 0.8889 32 0.6899 0.0 0.6899 0.8306
No log 0.9444 34 0.6329 -0.0159 0.6329 0.7955
No log 1.0 36 0.6195 -0.0159 0.6195 0.7871
No log 1.0556 38 0.6602 0.0189 0.6602 0.8126
No log 1.1111 40 0.8529 0.0823 0.8529 0.9235
No log 1.1667 42 0.7303 0.2410 0.7303 0.8546
No log 1.2222 44 0.6192 -0.0233 0.6192 0.7869
No log 1.2778 46 0.6132 -0.0233 0.6132 0.7831
No log 1.3333 48 0.6474 0.1724 0.6474 0.8046
No log 1.3889 50 0.6945 0.2727 0.6945 0.8334
No log 1.4444 52 0.5661 0.0725 0.5661 0.7524
No log 1.5 54 0.5648 0.1329 0.5648 0.7516
No log 1.5556 56 0.8069 0.2356 0.8069 0.8983
No log 1.6111 58 0.9849 0.1392 0.9849 0.9924
No log 1.6667 60 0.5963 0.2360 0.5963 0.7722
No log 1.7222 62 0.7808 0.2692 0.7808 0.8836
No log 1.7778 64 0.8629 0.2348 0.8629 0.9289
No log 1.8333 66 0.5103 0.2109 0.5103 0.7143
No log 1.8889 68 0.8237 0.2068 0.8237 0.9076
No log 1.9444 70 0.8621 0.1276 0.8621 0.9285
No log 2.0 72 0.7972 0.2208 0.7972 0.8928
No log 2.0556 74 0.5948 0.1605 0.5948 0.7712
No log 2.1111 76 0.5098 0.2432 0.5098 0.7140
No log 2.1667 78 0.9760 0.1680 0.9760 0.9879
No log 2.2222 80 0.9860 0.1385 0.9860 0.9930
No log 2.2778 82 0.5358 0.3333 0.5358 0.7320
No log 2.3333 84 0.5117 0.2663 0.5117 0.7153
No log 2.3889 86 0.5110 0.4227 0.5110 0.7148
No log 2.4444 88 0.6030 0.3706 0.6030 0.7765
No log 2.5 90 0.7309 0.2986 0.7309 0.8549
No log 2.5556 92 1.1166 0.1127 1.1166 1.0567
No log 2.6111 94 0.9337 0.1870 0.9337 0.9663
No log 2.6667 96 0.5569 0.3778 0.5569 0.7463
No log 2.7222 98 0.5615 0.3498 0.5615 0.7493
No log 2.7778 100 0.5565 0.4455 0.5565 0.7460
No log 2.8333 102 0.6659 0.3706 0.6659 0.8160
No log 2.8889 104 0.9556 0.2230 0.9556 0.9776
No log 2.9444 106 1.1222 0.1897 1.1222 1.0593
No log 3.0 108 0.6435 0.4338 0.6435 0.8022
No log 3.0556 110 0.7082 0.2511 0.7082 0.8415
No log 3.1111 112 0.6092 0.2308 0.6092 0.7805
No log 3.1667 114 0.8540 0.2000 0.8540 0.9241
No log 3.2222 116 1.7821 0.2038 1.7821 1.3350
No log 3.2778 118 1.4943 0.1957 1.4943 1.2224
No log 3.3333 120 0.6628 0.4667 0.6628 0.8141
No log 3.3889 122 0.5625 0.3433 0.5625 0.7500
No log 3.4444 124 0.5535 0.4526 0.5535 0.7440
No log 3.5 126 0.7457 0.2897 0.7457 0.8635
No log 3.5556 128 0.7102 0.2549 0.7102 0.8428
No log 3.6111 130 0.9883 0.1642 0.9883 0.9941
No log 3.6667 132 0.8894 0.2520 0.8894 0.9431
No log 3.7222 134 0.5907 0.3535 0.5907 0.7686
No log 3.7778 136 0.5636 0.4400 0.5636 0.7507
No log 3.8333 138 0.7261 0.2897 0.7261 0.8521
No log 3.8889 140 0.6773 0.2965 0.6773 0.8230
No log 3.9444 142 0.5429 0.3297 0.5429 0.7368
No log 4.0 144 0.4856 0.4595 0.4856 0.6969
No log 4.0556 146 0.4986 0.4098 0.4986 0.7061
No log 4.1111 148 0.5186 0.4222 0.5186 0.7202
No log 4.1667 150 0.5583 0.4162 0.5583 0.7472
No log 4.2222 152 0.6641 0.3398 0.6641 0.8149
No log 4.2778 154 0.7863 0.2982 0.7863 0.8867
No log 4.3333 156 0.7858 0.2982 0.7858 0.8864
No log 4.3889 158 0.5620 0.4227 0.5620 0.7496
No log 4.4444 160 0.5165 0.3367 0.5165 0.7187
No log 4.5 162 0.5487 0.4167 0.5487 0.7407
No log 4.5556 164 0.9769 0.2115 0.9769 0.9884
No log 4.6111 166 1.2966 0.1775 1.2966 1.1387
No log 4.6667 168 1.0073 0.2115 1.0073 1.0036
No log 4.7222 170 0.5254 0.4536 0.5254 0.7249
No log 4.7778 172 0.4944 0.2707 0.4944 0.7031
No log 4.8333 174 0.4859 0.4286 0.4859 0.6971
No log 4.8889 176 0.6701 0.3191 0.6701 0.8186
No log 4.9444 178 0.9140 0.2605 0.9140 0.9560
No log 5.0 180 0.8048 0.2381 0.8048 0.8971
No log 5.0556 182 0.5737 0.2727 0.5737 0.7574
No log 5.1111 184 0.5023 0.4091 0.5023 0.7088
No log 5.1667 186 0.5013 0.3439 0.5013 0.7080
No log 5.2222 188 0.5090 0.3846 0.5090 0.7134
No log 5.2778 190 0.5359 0.3333 0.5359 0.7321
No log 5.3333 192 0.7260 0.2759 0.7260 0.8521
No log 5.3889 194 0.7800 0.2743 0.7800 0.8832
No log 5.4444 196 0.5742 0.4167 0.5742 0.7577
No log 5.5 198 0.4878 0.3224 0.4878 0.6984
No log 5.5556 200 0.5045 0.3402 0.5045 0.7103
No log 5.6111 202 0.4890 0.4023 0.4890 0.6993
No log 5.6667 204 0.5806 0.3641 0.5806 0.7620
No log 5.7222 206 0.6121 0.3402 0.6121 0.7824
No log 5.7778 208 0.5336 0.3778 0.5336 0.7305
No log 5.8333 210 0.5174 0.3371 0.5174 0.7193
No log 5.8889 212 0.4847 0.3966 0.4847 0.6962
No log 5.9444 214 0.4855 0.3966 0.4855 0.6968
No log 6.0 216 0.5420 0.4694 0.5420 0.7362
No log 6.0556 218 0.6144 0.2919 0.6144 0.7839
No log 6.1111 220 0.6764 0.2621 0.6764 0.8224
No log 6.1667 222 0.5837 0.3402 0.5837 0.7640
No log 6.2222 224 0.4817 0.3407 0.4817 0.6940
No log 6.2778 226 0.4798 0.3814 0.4798 0.6927
No log 6.3333 228 0.4999 0.4098 0.4999 0.7071
No log 6.3889 230 0.5668 0.4112 0.5668 0.7529
No log 6.4444 232 0.6013 0.3623 0.6013 0.7754
No log 6.5 234 0.6635 0.2233 0.6635 0.8145
No log 6.5556 236 0.7752 0.2356 0.7752 0.8804
No log 6.6111 238 0.6964 0.2227 0.6964 0.8345
No log 6.6667 240 0.5318 0.4595 0.5318 0.7292
No log 6.7222 242 0.4786 0.4802 0.4786 0.6918
No log 6.7778 244 0.4706 0.4222 0.4706 0.6860
No log 6.8333 246 0.4775 0.4802 0.4775 0.6910
No log 6.8889 248 0.5567 0.4595 0.5567 0.7461
No log 6.9444 250 0.7362 0.2287 0.7362 0.8580
No log 7.0 252 0.8858 0.2667 0.8858 0.9412
No log 7.0556 254 0.8311 0.2275 0.8311 0.9117
No log 7.1111 256 0.6297 0.2621 0.6297 0.7935
No log 7.1667 258 0.4800 0.4725 0.4800 0.6928
No log 7.2222 260 0.4693 0.4098 0.4693 0.6851
No log 7.2778 262 0.4689 0.4802 0.4689 0.6848
No log 7.3333 264 0.4944 0.4348 0.4944 0.7032
No log 7.3889 266 0.5716 0.3892 0.5716 0.7560
No log 7.4444 268 0.6225 0.3488 0.6225 0.7890
No log 7.5 270 0.5785 0.3462 0.5785 0.7606
No log 7.5556 272 0.5235 0.4694 0.5235 0.7235
No log 7.6111 274 0.5214 0.4694 0.5214 0.7221
No log 7.6667 276 0.5183 0.4694 0.5183 0.7200
No log 7.7222 278 0.5113 0.4409 0.5113 0.7151
No log 7.7778 280 0.5009 0.4348 0.5009 0.7077
No log 7.8333 282 0.4969 0.4652 0.4969 0.7049
No log 7.8889 284 0.5140 0.4098 0.5140 0.7169
No log 7.9444 286 0.5527 0.3892 0.5527 0.7434
No log 8.0 288 0.5514 0.3892 0.5514 0.7426
No log 8.0556 290 0.5296 0.3939 0.5296 0.7277
No log 8.1111 292 0.5515 0.3892 0.5515 0.7426
No log 8.1667 294 0.5788 0.3267 0.5788 0.7608
No log 8.2222 296 0.5618 0.3892 0.5618 0.7496
No log 8.2778 298 0.5452 0.4229 0.5452 0.7384
No log 8.3333 300 0.5447 0.4286 0.5447 0.7381
No log 8.3889 302 0.5270 0.4286 0.5270 0.7260
No log 8.4444 304 0.5377 0.4286 0.5377 0.7333
No log 8.5 306 0.5425 0.3990 0.5425 0.7366
No log 8.5556 308 0.5417 0.3990 0.5417 0.7360
No log 8.6111 310 0.5434 0.3990 0.5434 0.7371
No log 8.6667 312 0.5242 0.4573 0.5242 0.7240
No log 8.7222 314 0.5017 0.5080 0.5017 0.7083
No log 8.7778 316 0.5041 0.5417 0.5041 0.7100
No log 8.8333 318 0.5060 0.5052 0.5060 0.7113
No log 8.8889 320 0.5229 0.4573 0.5229 0.7231
No log 8.9444 322 0.5427 0.3990 0.5427 0.7367
No log 9.0 324 0.5484 0.3990 0.5484 0.7406
No log 9.0556 326 0.5327 0.3990 0.5327 0.7298
No log 9.1111 328 0.5039 0.5052 0.5039 0.7098
No log 9.1667 330 0.4879 0.4725 0.4879 0.6985
No log 9.2222 332 0.4802 0.4802 0.4802 0.6929
No log 9.2778 334 0.4782 0.4802 0.4782 0.6915
No log 9.3333 336 0.4780 0.4802 0.4780 0.6914
No log 9.3889 338 0.4832 0.4725 0.4832 0.6951
No log 9.4444 340 0.4925 0.4783 0.4925 0.7018
No log 9.5 342 0.5046 0.5052 0.5046 0.7103
No log 9.5556 344 0.5175 0.4468 0.5175 0.7194
No log 9.6111 346 0.5299 0.4105 0.5299 0.7279
No log 9.6667 348 0.5413 0.4105 0.5413 0.7357
No log 9.7222 350 0.5423 0.4105 0.5423 0.7364
No log 9.7778 352 0.5411 0.4105 0.5411 0.7356
No log 9.8333 354 0.5404 0.4105 0.5404 0.7351
No log 9.8889 356 0.5376 0.4105 0.5376 0.7332
No log 9.9444 358 0.5354 0.4105 0.5354 0.7317
No log 10.0 360 0.5342 0.4105 0.5342 0.7309

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k7_task3_organization

Finetuned
(4023)
this model