ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k18_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8843
  • Qwk: 0.3258
  • Mse: 0.8843
  • Rmse: 0.9403

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0364 2 4.7824 0.0010 4.7824 2.1869
No log 0.0727 4 2.7949 -0.0452 2.7949 1.6718
No log 0.1091 6 2.1438 0.0094 2.1438 1.4642
No log 0.1455 8 2.7067 -0.0681 2.7067 1.6452
No log 0.1818 10 2.0233 -0.0233 2.0233 1.4224
No log 0.2182 12 1.5715 0.0018 1.5715 1.2536
No log 0.2545 14 1.4950 -0.0522 1.4950 1.2227
No log 0.2909 16 1.5672 -0.0534 1.5672 1.2519
No log 0.3273 18 1.5126 -0.0603 1.5126 1.2299
No log 0.3636 20 1.4470 -0.0603 1.4470 1.2029
No log 0.4 22 1.5509 0.0124 1.5509 1.2454
No log 0.4364 24 1.5838 0.0121 1.5838 1.2585
No log 0.4727 26 1.3843 0.1472 1.3843 1.1766
No log 0.5091 28 1.2289 0.0741 1.2289 1.1085
No log 0.5455 30 1.2237 0.0700 1.2237 1.1062
No log 0.5818 32 1.2593 0.0527 1.2593 1.1222
No log 0.6182 34 1.2912 0.0353 1.2912 1.1363
No log 0.6545 36 1.3308 0.0353 1.3308 1.1536
No log 0.6909 38 1.2902 -0.0043 1.2902 1.1359
No log 0.7273 40 1.2761 0.1011 1.2761 1.1297
No log 0.7636 42 1.2610 0.1579 1.2610 1.1229
No log 0.8 44 1.2323 0.1530 1.2323 1.1101
No log 0.8364 46 1.1637 0.1904 1.1637 1.0788
No log 0.8727 48 1.1364 0.2785 1.1364 1.0660
No log 0.9091 50 1.1317 0.2038 1.1317 1.0638
No log 0.9455 52 1.1248 0.1546 1.1248 1.0606
No log 0.9818 54 1.1400 0.1706 1.1400 1.0677
No log 1.0182 56 1.1534 0.1911 1.1534 1.0740
No log 1.0545 58 1.1542 0.1860 1.1542 1.0743
No log 1.0909 60 1.0527 0.2658 1.0527 1.0260
No log 1.1273 62 1.1391 0.3433 1.1391 1.0673
No log 1.1636 64 1.5475 0.1520 1.5475 1.2440
No log 1.2 66 1.7204 0.1618 1.7204 1.3116
No log 1.2364 68 1.6725 0.1140 1.6725 1.2933
No log 1.2727 70 1.4407 0.1743 1.4407 1.2003
No log 1.3091 72 1.2061 0.2191 1.2061 1.0982
No log 1.3455 74 1.0773 0.3908 1.0773 1.0379
No log 1.3818 76 1.0337 0.3612 1.0337 1.0167
No log 1.4182 78 1.0452 0.3793 1.0452 1.0224
No log 1.4545 80 1.0376 0.3567 1.0376 1.0186
No log 1.4909 82 1.0425 0.3090 1.0425 1.0210
No log 1.5273 84 1.0665 0.2590 1.0665 1.0327
No log 1.5636 86 1.0196 0.3021 1.0196 1.0097
No log 1.6 88 1.0450 0.3615 1.0450 1.0222
No log 1.6364 90 1.0477 0.3960 1.0477 1.0236
No log 1.6727 92 1.0849 0.3476 1.0849 1.0416
No log 1.7091 94 1.0528 0.4137 1.0528 1.0260
No log 1.7455 96 0.9493 0.4059 0.9493 0.9743
No log 1.7818 98 0.8759 0.4039 0.8759 0.9359
No log 1.8182 100 0.8620 0.4394 0.8620 0.9284
No log 1.8545 102 0.8678 0.4202 0.8678 0.9316
No log 1.8909 104 0.8925 0.3169 0.8925 0.9447
No log 1.9273 106 0.9197 0.3168 0.9197 0.9590
No log 1.9636 108 0.9250 0.3627 0.9250 0.9618
No log 2.0 110 0.8915 0.3237 0.8915 0.9442
No log 2.0364 112 0.9214 0.5 0.9214 0.9599
No log 2.0727 114 1.2593 0.4371 1.2593 1.1222
No log 2.1091 116 1.3732 0.2885 1.3732 1.1718
No log 2.1455 118 1.1124 0.4803 1.1124 1.0547
No log 2.1818 120 0.9339 0.4736 0.9339 0.9664
No log 2.2182 122 1.0178 0.4454 1.0178 1.0089
No log 2.2545 124 0.9630 0.4626 0.9630 0.9813
No log 2.2909 126 0.9311 0.3412 0.9311 0.9649
No log 2.3273 128 1.0026 0.3686 1.0026 1.0013
No log 2.3636 130 1.0006 0.2999 1.0006 1.0003
No log 2.4 132 1.0826 0.3398 1.0826 1.0405
No log 2.4364 134 1.0806 0.3068 1.0806 1.0395
No log 2.4727 136 0.9812 0.3412 0.9812 0.9905
No log 2.5091 138 0.9718 0.3695 0.9718 0.9858
No log 2.5455 140 0.9879 0.3654 0.9879 0.9939
No log 2.5818 142 1.0588 0.4167 1.0588 1.0290
No log 2.6182 144 1.0459 0.4220 1.0459 1.0227
No log 2.6545 146 0.9740 0.4132 0.9740 0.9869
No log 2.6909 148 0.8721 0.4002 0.8721 0.9338
No log 2.7273 150 0.8538 0.3920 0.8538 0.9240
No log 2.7636 152 0.8603 0.4159 0.8603 0.9275
No log 2.8 154 0.8722 0.4316 0.8722 0.9339
No log 2.8364 156 0.9145 0.4019 0.9145 0.9563
No log 2.8727 158 0.8906 0.3694 0.8906 0.9437
No log 2.9091 160 0.8884 0.3914 0.8884 0.9425
No log 2.9455 162 0.9723 0.4136 0.9723 0.9861
No log 2.9818 164 0.9800 0.4136 0.9800 0.9899
No log 3.0182 166 0.9571 0.4408 0.9571 0.9783
No log 3.0545 168 0.8827 0.4388 0.8827 0.9395
No log 3.0909 170 0.9338 0.3986 0.9338 0.9663
No log 3.1273 172 0.9264 0.3979 0.9264 0.9625
No log 3.1636 174 0.8948 0.4013 0.8948 0.9459
No log 3.2 176 0.8484 0.3873 0.8484 0.9211
No log 3.2364 178 0.8669 0.5060 0.8669 0.9311
No log 3.2727 180 0.9149 0.4974 0.9149 0.9565
No log 3.3091 182 0.8992 0.4541 0.8992 0.9483
No log 3.3455 184 0.8826 0.2813 0.8826 0.9395
No log 3.3818 186 0.9026 0.3289 0.9026 0.9501
No log 3.4182 188 0.9069 0.3352 0.9069 0.9523
No log 3.4545 190 0.9107 0.4252 0.9107 0.9543
No log 3.4909 192 1.0391 0.4155 1.0391 1.0194
No log 3.5273 194 1.1321 0.4106 1.1321 1.0640
No log 3.5636 196 1.0243 0.4991 1.0243 1.0121
No log 3.6 198 0.9635 0.3960 0.9635 0.9816
No log 3.6364 200 0.9562 0.3392 0.9562 0.9778
No log 3.6727 202 0.9644 0.3478 0.9644 0.9820
No log 3.7091 204 0.9347 0.2692 0.9347 0.9668
No log 3.7455 206 0.9319 0.3299 0.9319 0.9653
No log 3.7818 208 0.9468 0.4411 0.9468 0.9730
No log 3.8182 210 0.9296 0.3250 0.9296 0.9642
No log 3.8545 212 0.9403 0.2843 0.9403 0.9697
No log 3.8909 214 0.9374 0.2843 0.9374 0.9682
No log 3.9273 216 0.9404 0.3195 0.9404 0.9698
No log 3.9636 218 0.9602 0.2440 0.9602 0.9799
No log 4.0 220 0.9907 0.1927 0.9907 0.9953
No log 4.0364 222 1.0041 0.1717 1.0041 1.0020
No log 4.0727 224 0.9853 0.1605 0.9853 0.9926
No log 4.1091 226 0.9975 0.1406 0.9975 0.9988
No log 4.1455 228 0.9893 0.1546 0.9893 0.9946
No log 4.1818 230 0.9938 0.1546 0.9938 0.9969
No log 4.2182 232 0.9687 0.2569 0.9687 0.9842
No log 4.2545 234 0.9625 0.1875 0.9625 0.9810
No log 4.2909 236 0.9398 0.2729 0.9398 0.9694
No log 4.3273 238 0.9210 0.3431 0.9210 0.9597
No log 4.3636 240 0.9102 0.3431 0.9102 0.9541
No log 4.4 242 0.9057 0.3169 0.9057 0.9517
No log 4.4364 244 0.8963 0.3021 0.8963 0.9467
No log 4.4727 246 0.8902 0.3779 0.8902 0.9435
No log 4.5091 248 0.9710 0.4344 0.9710 0.9854
No log 4.5455 250 1.0038 0.4739 1.0038 1.0019
No log 4.5818 252 0.9293 0.3451 0.9293 0.9640
No log 4.6182 254 0.8947 0.3779 0.8947 0.9459
No log 4.6545 256 0.8860 0.2972 0.8860 0.9413
No log 4.6909 258 0.8659 0.3564 0.8659 0.9305
No log 4.7273 260 0.8653 0.4180 0.8653 0.9302
No log 4.7636 262 0.9072 0.4261 0.9072 0.9525
No log 4.8 264 0.8839 0.3866 0.8839 0.9402
No log 4.8364 266 0.8665 0.3629 0.8665 0.9309
No log 4.8727 268 0.8676 0.3095 0.8676 0.9314
No log 4.9091 270 0.8764 0.3223 0.8764 0.9361
No log 4.9455 272 0.8615 0.3318 0.8615 0.9281
No log 4.9818 274 0.8593 0.3318 0.8593 0.9270
No log 5.0182 276 0.8689 0.3250 0.8689 0.9321
No log 5.0545 278 0.9033 0.2917 0.9033 0.9504
No log 5.0909 280 0.9626 0.3000 0.9626 0.9811
No log 5.1273 282 0.8926 0.3777 0.8926 0.9448
No log 5.1636 284 0.8580 0.3298 0.8580 0.9263
No log 5.2 286 0.8371 0.3890 0.8371 0.9149
No log 5.2364 288 0.8275 0.4258 0.8275 0.9097
No log 5.2727 290 0.8452 0.4527 0.8452 0.9193
No log 5.3091 292 0.8813 0.3943 0.8813 0.9388
No log 5.3455 294 0.8673 0.3920 0.8673 0.9313
No log 5.3818 296 0.8699 0.4991 0.8699 0.9327
No log 5.4182 298 0.8740 0.5023 0.8740 0.9349
No log 5.4545 300 0.8416 0.4234 0.8416 0.9174
No log 5.4909 302 0.8373 0.4181 0.8373 0.9150
No log 5.5273 304 0.8391 0.4700 0.8391 0.9160
No log 5.5636 306 0.7779 0.4196 0.7779 0.8820
No log 5.6 308 0.8144 0.5247 0.8144 0.9024
No log 5.6364 310 1.0304 0.4862 1.0304 1.0151
No log 5.6727 312 1.0825 0.4104 1.0825 1.0404
No log 5.7091 314 0.8741 0.4473 0.8741 0.9349
No log 5.7455 316 0.7851 0.4369 0.7851 0.8861
No log 5.7818 318 0.8601 0.4732 0.8601 0.9274
No log 5.8182 320 0.8105 0.4069 0.8105 0.9003
No log 5.8545 322 0.7778 0.5195 0.7778 0.8819
No log 5.8909 324 0.7779 0.4898 0.7779 0.8820
No log 5.9273 326 0.7905 0.4343 0.7905 0.8891
No log 5.9636 328 0.8911 0.4339 0.8911 0.9440
No log 6.0 330 0.9738 0.4304 0.9738 0.9868
No log 6.0364 332 0.9113 0.4130 0.9113 0.9546
No log 6.0727 334 0.8382 0.4181 0.8382 0.9155
No log 6.1091 336 0.8554 0.4364 0.8554 0.9249
No log 6.1455 338 0.8614 0.4119 0.8614 0.9281
No log 6.1818 340 0.9099 0.2969 0.9099 0.9539
No log 6.2182 342 0.9356 0.2969 0.9356 0.9672
No log 6.2545 344 0.8917 0.2834 0.8917 0.9443
No log 6.2909 346 0.8812 0.3196 0.8812 0.9387
No log 6.3273 348 0.8405 0.3943 0.8405 0.9168
No log 6.3636 350 0.8214 0.3695 0.8214 0.9063
No log 6.4 352 0.8174 0.4159 0.8174 0.9041
No log 6.4364 354 0.8499 0.4840 0.8499 0.9219
No log 6.4727 356 0.9748 0.5014 0.9748 0.9873
No log 6.5091 358 1.0998 0.4186 1.0998 1.0487
No log 6.5455 360 1.0189 0.3424 1.0189 1.0094
No log 6.5818 362 0.9137 0.3926 0.9137 0.9559
No log 6.6182 364 0.9012 0.3926 0.9012 0.9493
No log 6.6545 366 0.9600 0.5173 0.9600 0.9798
No log 6.6909 368 1.1336 0.4371 1.1336 1.0647
No log 6.7273 370 1.0746 0.4963 1.0746 1.0366
No log 6.7636 372 0.8433 0.4956 0.8433 0.9183
No log 6.8 374 0.7960 0.5380 0.7960 0.8922
No log 6.8364 376 0.8238 0.4666 0.8238 0.9076
No log 6.8727 378 0.8279 0.4676 0.8279 0.9099
No log 6.9091 380 0.8967 0.3989 0.8967 0.9469
No log 6.9455 382 0.9110 0.3762 0.9110 0.9545
No log 6.9818 384 0.9094 0.3762 0.9094 0.9536
No log 7.0182 386 0.9387 0.3394 0.9387 0.9689
No log 7.0545 388 0.8888 0.3989 0.8888 0.9428
No log 7.0909 390 0.8529 0.4539 0.8529 0.9235
No log 7.1273 392 0.8219 0.4482 0.8219 0.9066
No log 7.1636 394 0.8492 0.4982 0.8492 0.9215
No log 7.2 396 0.9089 0.4003 0.9089 0.9534
No log 7.2364 398 0.8781 0.4351 0.8781 0.9371
No log 7.2727 400 0.7975 0.5062 0.7975 0.8930
No log 7.3091 402 0.7891 0.5044 0.7891 0.8883
No log 7.3455 404 0.8011 0.5112 0.8011 0.8950
No log 7.3818 406 0.8364 0.4402 0.8364 0.9145
No log 7.4182 408 0.8562 0.4264 0.8562 0.9253
No log 7.4545 410 0.8272 0.4402 0.8272 0.9095
No log 7.4909 412 0.7772 0.5027 0.7772 0.8816
No log 7.5273 414 0.7903 0.4719 0.7903 0.8890
No log 7.5636 416 0.7856 0.4912 0.7856 0.8863
No log 7.6 418 0.7911 0.4961 0.7911 0.8894
No log 7.6364 420 0.7620 0.5312 0.7620 0.8729
No log 7.6727 422 0.7722 0.5230 0.7722 0.8787
No log 7.7091 424 0.9032 0.4958 0.9032 0.9504
No log 7.7455 426 0.9073 0.4958 0.9073 0.9525
No log 7.7818 428 0.8033 0.5029 0.8033 0.8963
No log 7.8182 430 0.7719 0.5305 0.7719 0.8786
No log 7.8545 432 0.7582 0.5536 0.7582 0.8708
No log 7.8909 434 0.7659 0.5324 0.7659 0.8751
No log 7.9273 436 0.7777 0.5324 0.7777 0.8819
No log 7.9636 438 0.7516 0.5308 0.7516 0.8670
No log 8.0 440 0.7435 0.5580 0.7435 0.8622
No log 8.0364 442 0.7507 0.5948 0.7507 0.8664
No log 8.0727 444 0.7763 0.4826 0.7763 0.8811
No log 8.1091 446 0.7957 0.4180 0.7957 0.8920
No log 8.1455 448 0.8458 0.3804 0.8458 0.9197
No log 8.1818 450 0.8635 0.2941 0.8635 0.9292
No log 8.2182 452 0.8392 0.3151 0.8392 0.9161
No log 8.2545 454 0.8358 0.3100 0.8358 0.9142
No log 8.2909 456 0.8319 0.3760 0.8319 0.9121
No log 8.3273 458 0.7946 0.4243 0.7946 0.8914
No log 8.3636 460 0.7736 0.4282 0.7736 0.8795
No log 8.4 462 0.7533 0.5012 0.7533 0.8679
No log 8.4364 464 0.7491 0.5688 0.7491 0.8655
No log 8.4727 466 0.7626 0.5959 0.7626 0.8733
No log 8.5091 468 0.7626 0.5959 0.7626 0.8733
No log 8.5455 470 0.7249 0.5283 0.7249 0.8514
No log 8.5818 472 0.7529 0.5059 0.7529 0.8677
No log 8.6182 474 0.7566 0.4724 0.7566 0.8698
No log 8.6545 476 0.7681 0.4321 0.7681 0.8764
No log 8.6909 478 0.7821 0.3734 0.7821 0.8843
No log 8.7273 480 0.8013 0.3779 0.8013 0.8951
No log 8.7636 482 0.8108 0.4220 0.8108 0.9004
No log 8.8 484 0.8032 0.4220 0.8032 0.8962
No log 8.8364 486 0.8011 0.4385 0.8011 0.8950
No log 8.8727 488 0.8913 0.4720 0.8913 0.9441
No log 8.9091 490 0.9431 0.4894 0.9431 0.9711
No log 8.9455 492 0.8564 0.4532 0.8564 0.9254
No log 8.9818 494 0.7779 0.4119 0.7779 0.8820
No log 9.0182 496 0.7767 0.4941 0.7767 0.8813
No log 9.0545 498 0.7771 0.4570 0.7771 0.8816
0.4115 9.0909 500 0.7892 0.4180 0.7892 0.8884
0.4115 9.1273 502 0.9029 0.4262 0.9029 0.9502
0.4115 9.1636 504 0.9866 0.4435 0.9866 0.9933
0.4115 9.2 506 0.9508 0.3223 0.9508 0.9751
0.4115 9.2364 508 0.8878 0.3583 0.8878 0.9422
0.4115 9.2727 510 0.8541 0.3629 0.8541 0.9242
0.4115 9.3091 512 0.8543 0.4139 0.8543 0.9243
0.4115 9.3455 514 0.8843 0.3258 0.8843 0.9403

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k18_task2_organization

Finetuned
(4023)
this model