ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k9_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9318
  • Qwk: 0.5282
  • Mse: 0.9318
  • Rmse: 0.9653

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0417 2 4.0348 0.0162 4.0348 2.0087
No log 0.0833 4 2.0396 0.0457 2.0396 1.4281
No log 0.125 6 1.0496 0.0569 1.0496 1.0245
No log 0.1667 8 0.7748 0.1041 0.7748 0.8802
No log 0.2083 10 0.6987 0.2638 0.6987 0.8359
No log 0.25 12 0.7123 0.2333 0.7123 0.8440
No log 0.2917 14 0.7193 0.1519 0.7193 0.8481
No log 0.3333 16 0.7155 0.1630 0.7155 0.8459
No log 0.375 18 0.6941 0.1434 0.6941 0.8331
No log 0.4167 20 0.6740 0.2256 0.6740 0.8210
No log 0.4583 22 0.9106 0.3964 0.9106 0.9543
No log 0.5 24 1.1471 0.3330 1.1471 1.0710
No log 0.5417 26 0.8916 0.3824 0.8916 0.9443
No log 0.5833 28 0.6764 0.3578 0.6764 0.8225
No log 0.625 30 0.5834 0.4454 0.5834 0.7638
No log 0.6667 32 0.6095 0.4686 0.6095 0.7807
No log 0.7083 34 0.6460 0.4782 0.6460 0.8038
No log 0.75 36 0.8466 0.4301 0.8466 0.9201
No log 0.7917 38 0.9294 0.3699 0.9294 0.9640
No log 0.8333 40 0.7949 0.4779 0.7949 0.8916
No log 0.875 42 0.6658 0.5171 0.6658 0.8160
No log 0.9167 44 0.6442 0.5525 0.6442 0.8026
No log 0.9583 46 0.6615 0.5677 0.6615 0.8133
No log 1.0 48 0.7146 0.5550 0.7146 0.8454
No log 1.0417 50 0.8115 0.5069 0.8115 0.9008
No log 1.0833 52 0.9127 0.5241 0.9127 0.9553
No log 1.125 54 0.8788 0.5164 0.8788 0.9374
No log 1.1667 56 0.8924 0.5493 0.8924 0.9447
No log 1.2083 58 0.9816 0.5029 0.9816 0.9908
No log 1.25 60 0.9885 0.5063 0.9885 0.9942
No log 1.2917 62 1.0158 0.5087 1.0158 1.0079
No log 1.3333 64 1.0444 0.5171 1.0444 1.0220
No log 1.375 66 0.9860 0.5236 0.9860 0.9930
No log 1.4167 68 0.9476 0.5433 0.9476 0.9734
No log 1.4583 70 0.9291 0.5300 0.9291 0.9639
No log 1.5 72 0.9597 0.5009 0.9597 0.9796
No log 1.5417 74 0.8636 0.5362 0.8636 0.9293
No log 1.5833 76 0.9222 0.5265 0.9222 0.9603
No log 1.625 78 0.9437 0.5516 0.9437 0.9715
No log 1.6667 80 0.9920 0.4886 0.9920 0.9960
No log 1.7083 82 0.8668 0.5609 0.8668 0.9310
No log 1.75 84 0.8182 0.5658 0.8182 0.9045
No log 1.7917 86 0.8071 0.5454 0.8071 0.8984
No log 1.8333 88 0.8645 0.5204 0.8645 0.9298
No log 1.875 90 0.9073 0.5388 0.9073 0.9525
No log 1.9167 92 0.9970 0.4816 0.9970 0.9985
No log 1.9583 94 1.2769 0.4245 1.2769 1.1300
No log 2.0 96 1.2736 0.4192 1.2736 1.1286
No log 2.0417 98 1.1122 0.5016 1.1122 1.0546
No log 2.0833 100 1.2060 0.4396 1.2060 1.0982
No log 2.125 102 1.1800 0.4281 1.1800 1.0863
No log 2.1667 104 0.9291 0.5173 0.9291 0.9639
No log 2.2083 106 0.9896 0.4722 0.9896 0.9948
No log 2.25 108 1.0932 0.3827 1.0932 1.0455
No log 2.2917 110 0.8950 0.5273 0.8950 0.9461
No log 2.3333 112 0.7190 0.5634 0.7190 0.8480
No log 2.375 114 0.9544 0.4926 0.9544 0.9770
No log 2.4167 116 1.0471 0.4611 1.0471 1.0233
No log 2.4583 118 0.8133 0.4728 0.8133 0.9018
No log 2.5 120 0.6416 0.5642 0.6416 0.8010
No log 2.5417 122 0.7410 0.5223 0.7410 0.8608
No log 2.5833 124 0.8141 0.5418 0.8141 0.9023
No log 2.625 126 0.7749 0.5503 0.7749 0.8803
No log 2.6667 128 0.8801 0.5360 0.8801 0.9381
No log 2.7083 130 1.0722 0.4684 1.0722 1.0355
No log 2.75 132 1.0855 0.4310 1.0855 1.0419
No log 2.7917 134 0.9938 0.5528 0.9938 0.9969
No log 2.8333 136 0.9915 0.5354 0.9915 0.9957
No log 2.875 138 0.9943 0.5410 0.9943 0.9972
No log 2.9167 140 0.9878 0.5418 0.9878 0.9939
No log 2.9583 142 1.0240 0.4996 1.0240 1.0119
No log 3.0 144 1.1516 0.4402 1.1516 1.0731
No log 3.0417 146 1.0416 0.4502 1.0416 1.0206
No log 3.0833 148 0.8655 0.5333 0.8655 0.9303
No log 3.125 150 0.8445 0.5716 0.8445 0.9189
No log 3.1667 152 0.8740 0.5841 0.8740 0.9349
No log 3.2083 154 0.8307 0.5396 0.8307 0.9115
No log 3.25 156 0.8666 0.5346 0.8666 0.9309
No log 3.2917 158 0.8833 0.5085 0.8833 0.9398
No log 3.3333 160 0.8244 0.5287 0.8244 0.9080
No log 3.375 162 0.7674 0.5123 0.7674 0.8760
No log 3.4167 164 0.7628 0.5357 0.7628 0.8734
No log 3.4583 166 0.8346 0.5586 0.8346 0.9135
No log 3.5 168 0.8380 0.5545 0.8380 0.9154
No log 3.5417 170 0.8111 0.5450 0.8111 0.9006
No log 3.5833 172 0.9048 0.4904 0.9048 0.9512
No log 3.625 174 1.1355 0.4686 1.1355 1.0656
No log 3.6667 176 1.1607 0.4637 1.1607 1.0774
No log 3.7083 178 0.9899 0.4650 0.9899 0.9949
No log 3.75 180 0.9066 0.5308 0.9066 0.9521
No log 3.7917 182 1.0682 0.4512 1.0682 1.0335
No log 3.8333 184 1.1159 0.4372 1.1159 1.0564
No log 3.875 186 1.0066 0.5487 1.0066 1.0033
No log 3.9167 188 0.9430 0.5156 0.9430 0.9711
No log 3.9583 190 1.0390 0.4454 1.0390 1.0193
No log 4.0 192 1.0383 0.4760 1.0383 1.0189
No log 4.0417 194 0.9229 0.4948 0.9229 0.9607
No log 4.0833 196 0.8983 0.5592 0.8983 0.9478
No log 4.125 198 0.9916 0.4691 0.9916 0.9958
No log 4.1667 200 1.0569 0.4623 1.0569 1.0280
No log 4.2083 202 1.0388 0.4889 1.0388 1.0192
No log 4.25 204 1.0839 0.4959 1.0839 1.0411
No log 4.2917 206 1.1379 0.4730 1.1379 1.0667
No log 4.3333 208 1.0780 0.4806 1.0780 1.0383
No log 4.375 210 0.9939 0.5484 0.9939 0.9969
No log 4.4167 212 0.9684 0.5139 0.9684 0.9841
No log 4.4583 214 0.9477 0.5484 0.9477 0.9735
No log 4.5 216 0.9493 0.5611 0.9493 0.9743
No log 4.5417 218 0.9868 0.5643 0.9868 0.9934
No log 4.5833 220 1.0550 0.4964 1.0550 1.0272
No log 4.625 222 1.0898 0.4897 1.0898 1.0439
No log 4.6667 224 1.1110 0.5248 1.1110 1.0540
No log 4.7083 226 1.1991 0.4970 1.1991 1.0950
No log 4.75 228 1.3104 0.4704 1.3104 1.1447
No log 4.7917 230 1.2896 0.4838 1.2896 1.1356
No log 4.8333 232 1.1445 0.5068 1.1445 1.0698
No log 4.875 234 1.0250 0.5322 1.0250 1.0124
No log 4.9167 236 0.9961 0.5320 0.9961 0.9980
No log 4.9583 238 0.9883 0.5320 0.9883 0.9942
No log 5.0 240 0.9789 0.5436 0.9789 0.9894
No log 5.0417 242 0.9521 0.5521 0.9521 0.9758
No log 5.0833 244 0.9731 0.5582 0.9731 0.9865
No log 5.125 246 1.0132 0.5531 1.0132 1.0066
No log 5.1667 248 1.0260 0.5541 1.0260 1.0129
No log 5.2083 250 1.0277 0.5470 1.0277 1.0138
No log 5.25 252 0.9771 0.5706 0.9771 0.9885
No log 5.2917 254 0.9015 0.5776 0.9015 0.9495
No log 5.3333 256 0.8483 0.5566 0.8483 0.9211
No log 5.375 258 0.8015 0.5552 0.8015 0.8953
No log 5.4167 260 0.8032 0.5707 0.8032 0.8962
No log 5.4583 262 0.8541 0.5428 0.8541 0.9242
No log 5.5 264 0.8453 0.5391 0.8453 0.9194
No log 5.5417 266 0.8668 0.5739 0.8668 0.9310
No log 5.5833 268 0.9574 0.5168 0.9574 0.9785
No log 5.625 270 1.0889 0.4472 1.0889 1.0435
No log 5.6667 272 1.1682 0.4550 1.1682 1.0808
No log 5.7083 274 1.1324 0.4562 1.1324 1.0642
No log 5.75 276 1.0470 0.4934 1.0470 1.0232
No log 5.7917 278 0.9578 0.5346 0.9578 0.9787
No log 5.8333 280 0.9010 0.5540 0.9010 0.9492
No log 5.875 282 0.8599 0.5553 0.8599 0.9273
No log 5.9167 284 0.8434 0.5367 0.8434 0.9183
No log 5.9583 286 0.8405 0.5389 0.8405 0.9168
No log 6.0 288 0.8728 0.5191 0.8728 0.9342
No log 6.0417 290 0.9425 0.4630 0.9425 0.9708
No log 6.0833 292 0.9375 0.4797 0.9375 0.9683
No log 6.125 294 0.8818 0.4960 0.8818 0.9390
No log 6.1667 296 0.8250 0.5549 0.8250 0.9083
No log 6.2083 298 0.8338 0.5502 0.8338 0.9131
No log 6.25 300 0.8615 0.5259 0.8615 0.9282
No log 6.2917 302 0.8851 0.5566 0.8851 0.9408
No log 6.3333 304 0.9089 0.5413 0.9089 0.9534
No log 6.375 306 0.9529 0.5225 0.9529 0.9762
No log 6.4167 308 0.9878 0.5210 0.9878 0.9939
No log 6.4583 310 0.9820 0.5314 0.9820 0.9909
No log 6.5 312 0.9680 0.5311 0.9680 0.9839
No log 6.5417 314 0.9139 0.5310 0.9139 0.9560
No log 6.5833 316 0.8855 0.5339 0.8855 0.9410
No log 6.625 318 0.8703 0.5449 0.8703 0.9329
No log 6.6667 320 0.8538 0.5450 0.8538 0.9240
No log 6.7083 322 0.8498 0.5449 0.8498 0.9218
No log 6.75 324 0.8609 0.5368 0.8609 0.9278
No log 6.7917 326 0.8792 0.5288 0.8792 0.9377
No log 6.8333 328 0.9188 0.5064 0.9188 0.9586
No log 6.875 330 0.9769 0.4975 0.9769 0.9884
No log 6.9167 332 0.9846 0.4981 0.9846 0.9923
No log 6.9583 334 0.9626 0.5044 0.9626 0.9811
No log 7.0 336 0.9565 0.5095 0.9565 0.9780
No log 7.0417 338 0.9441 0.5655 0.9441 0.9716
No log 7.0833 340 0.9224 0.5564 0.9224 0.9604
No log 7.125 342 0.9043 0.5527 0.9043 0.9510
No log 7.1667 344 0.8970 0.5262 0.8970 0.9471
No log 7.2083 346 0.9097 0.5202 0.9097 0.9538
No log 7.25 348 0.9364 0.4868 0.9364 0.9677
No log 7.2917 350 0.9263 0.4868 0.9263 0.9624
No log 7.3333 352 0.8981 0.5252 0.8981 0.9477
No log 7.375 354 0.8996 0.5195 0.8996 0.9485
No log 7.4167 356 0.9235 0.5195 0.9235 0.9610
No log 7.4583 358 0.9569 0.4990 0.9569 0.9782
No log 7.5 360 0.9655 0.4929 0.9655 0.9826
No log 7.5417 362 0.9624 0.4917 0.9624 0.9810
No log 7.5833 364 0.9317 0.5138 0.9317 0.9653
No log 7.625 366 0.9088 0.5560 0.9088 0.9533
No log 7.6667 368 0.8907 0.5336 0.8907 0.9438
No log 7.7083 370 0.8872 0.5438 0.8872 0.9419
No log 7.75 372 0.8847 0.5437 0.8847 0.9406
No log 7.7917 374 0.8871 0.5438 0.8871 0.9418
No log 7.8333 376 0.8830 0.5323 0.8830 0.9397
No log 7.875 378 0.8955 0.5152 0.8955 0.9463
No log 7.9167 380 0.9125 0.5342 0.9125 0.9552
No log 7.9583 382 0.9171 0.5193 0.9171 0.9577
No log 8.0 384 0.9138 0.5120 0.9138 0.9559
No log 8.0417 386 0.9066 0.5478 0.9066 0.9521
No log 8.0833 388 0.9079 0.5468 0.9079 0.9528
No log 8.125 390 0.9144 0.5468 0.9144 0.9563
No log 8.1667 392 0.9134 0.5492 0.9134 0.9557
No log 8.2083 394 0.9157 0.5417 0.9157 0.9569
No log 8.25 396 0.9109 0.5286 0.9109 0.9544
No log 8.2917 398 0.9151 0.5404 0.9151 0.9566
No log 8.3333 400 0.9170 0.5344 0.9170 0.9576
No log 8.375 402 0.9034 0.5488 0.9034 0.9505
No log 8.4167 404 0.8962 0.5391 0.8962 0.9467
No log 8.4583 406 0.8822 0.5472 0.8822 0.9392
No log 8.5 408 0.8680 0.5373 0.8680 0.9317
No log 8.5417 410 0.8589 0.5331 0.8589 0.9268
No log 8.5833 412 0.8560 0.5331 0.8560 0.9252
No log 8.625 414 0.8528 0.5331 0.8528 0.9235
No log 8.6667 416 0.8511 0.5331 0.8511 0.9225
No log 8.7083 418 0.8471 0.5411 0.8471 0.9204
No log 8.75 420 0.8465 0.5538 0.8465 0.9201
No log 8.7917 422 0.8530 0.5538 0.8530 0.9236
No log 8.8333 424 0.8569 0.5538 0.8569 0.9257
No log 8.875 426 0.8628 0.5588 0.8628 0.9288
No log 8.9167 428 0.8739 0.5485 0.8739 0.9348
No log 8.9583 430 0.8964 0.5347 0.8964 0.9468
No log 9.0 432 0.9347 0.5089 0.9347 0.9668
No log 9.0417 434 0.9628 0.5024 0.9628 0.9812
No log 9.0833 436 0.9829 0.4967 0.9829 0.9914
No log 9.125 438 0.9974 0.4955 0.9974 0.9987
No log 9.1667 440 0.9927 0.4911 0.9927 0.9963
No log 9.2083 442 0.9852 0.5024 0.9852 0.9926
No log 9.25 444 0.9705 0.4889 0.9705 0.9851
No log 9.2917 446 0.9526 0.5173 0.9526 0.9760
No log 9.3333 448 0.9452 0.5045 0.9452 0.9722
No log 9.375 450 0.9361 0.5454 0.9361 0.9675
No log 9.4167 452 0.9312 0.5262 0.9312 0.9650
No log 9.4583 454 0.9280 0.5326 0.9280 0.9633
No log 9.5 456 0.9281 0.5362 0.9281 0.9634
No log 9.5417 458 0.9313 0.5454 0.9313 0.9650
No log 9.5833 460 0.9351 0.5282 0.9351 0.9670
No log 9.625 462 0.9381 0.5170 0.9381 0.9685
No log 9.6667 464 0.9395 0.5110 0.9395 0.9693
No log 9.7083 466 0.9395 0.5110 0.9395 0.9693
No log 9.75 468 0.9384 0.5170 0.9384 0.9687
No log 9.7917 470 0.9364 0.5170 0.9364 0.9677
No log 9.8333 472 0.9358 0.5170 0.9358 0.9674
No log 9.875 474 0.9347 0.5170 0.9347 0.9668
No log 9.9167 476 0.9332 0.5282 0.9332 0.9660
No log 9.9583 478 0.9321 0.5282 0.9321 0.9655
No log 10.0 480 0.9318 0.5282 0.9318 0.9653

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k9_task2_organization

Finetuned
(4023)
this model