ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k9_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8955
  • Qwk: 0.5255
  • Mse: 0.8955
  • Rmse: 0.9463

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0392 2 3.9326 0.0107 3.9326 1.9831
No log 0.0784 4 1.7594 0.1289 1.7594 1.3264
No log 0.1176 6 0.8643 0.1583 0.8643 0.9297
No log 0.1569 8 0.8245 0.1679 0.8245 0.9080
No log 0.1961 10 1.4197 0.1594 1.4197 1.1915
No log 0.2353 12 1.3437 0.1978 1.3437 1.1592
No log 0.2745 14 0.9190 0.2023 0.9190 0.9586
No log 0.3137 16 1.0801 0.2387 1.0801 1.0393
No log 0.3529 18 0.8076 0.2273 0.8076 0.8987
No log 0.3922 20 0.6001 0.2793 0.6001 0.7747
No log 0.4314 22 0.5741 0.2917 0.5741 0.7577
No log 0.4706 24 0.6721 0.3758 0.6721 0.8198
No log 0.5098 26 0.7048 0.3796 0.7048 0.8395
No log 0.5490 28 0.6240 0.4502 0.6240 0.7900
No log 0.5882 30 0.5652 0.4388 0.5652 0.7518
No log 0.6275 32 0.5825 0.4758 0.5825 0.7632
No log 0.6667 34 0.7838 0.4488 0.7838 0.8853
No log 0.7059 36 0.8217 0.4626 0.8217 0.9065
No log 0.7451 38 0.6790 0.5171 0.6790 0.8240
No log 0.7843 40 0.8307 0.4766 0.8307 0.9114
No log 0.8235 42 1.3523 0.2529 1.3523 1.1629
No log 0.8627 44 1.1371 0.3219 1.1371 1.0663
No log 0.9020 46 0.7247 0.5253 0.7247 0.8513
No log 0.9412 48 0.7632 0.5675 0.7632 0.8736
No log 0.9804 50 0.7451 0.5197 0.7451 0.8632
No log 1.0196 52 0.9024 0.4224 0.9024 0.9500
No log 1.0588 54 0.7789 0.5006 0.7789 0.8826
No log 1.0980 56 0.7431 0.5281 0.7431 0.8620
No log 1.1373 58 0.8569 0.5157 0.8569 0.9257
No log 1.1765 60 0.9439 0.4609 0.9439 0.9716
No log 1.2157 62 0.8141 0.5285 0.8141 0.9022
No log 1.2549 64 1.0059 0.4212 1.0059 1.0029
No log 1.2941 66 0.9391 0.4666 0.9391 0.9690
No log 1.3333 68 0.8543 0.5327 0.8543 0.9243
No log 1.3725 70 0.9626 0.5029 0.9626 0.9811
No log 1.4118 72 0.9282 0.4784 0.9282 0.9634
No log 1.4510 74 0.7656 0.5249 0.7656 0.8750
No log 1.4902 76 0.7762 0.4943 0.7762 0.8810
No log 1.5294 78 0.7449 0.4903 0.7449 0.8631
No log 1.5686 80 0.7536 0.4893 0.7536 0.8681
No log 1.6078 82 0.8231 0.5814 0.8231 0.9072
No log 1.6471 84 0.9383 0.4972 0.9383 0.9687
No log 1.6863 86 0.9541 0.5406 0.9541 0.9768
No log 1.7255 88 0.9804 0.5427 0.9804 0.9902
No log 1.7647 90 1.1314 0.4371 1.1314 1.0637
No log 1.8039 92 1.3730 0.3958 1.3730 1.1717
No log 1.8431 94 1.2803 0.4442 1.2803 1.1315
No log 1.8824 96 1.1683 0.5026 1.1683 1.0809
No log 1.9216 98 1.1196 0.5129 1.1196 1.0581
No log 1.9608 100 1.1056 0.4682 1.1056 1.0515
No log 2.0 102 1.2410 0.4059 1.2410 1.1140
No log 2.0392 104 1.2126 0.3733 1.2126 1.1012
No log 2.0784 106 0.9092 0.5106 0.9092 0.9535
No log 2.1176 108 0.7451 0.5294 0.7451 0.8632
No log 2.1569 110 0.7006 0.5595 0.7006 0.8370
No log 2.1961 112 0.6808 0.5237 0.6808 0.8251
No log 2.2353 114 0.7468 0.5115 0.7468 0.8642
No log 2.2745 116 0.8129 0.5323 0.8129 0.9016
No log 2.3137 118 0.8708 0.5356 0.8708 0.9332
No log 2.3529 120 0.9206 0.5127 0.9206 0.9595
No log 2.3922 122 0.8564 0.5377 0.8564 0.9254
No log 2.4314 124 0.8903 0.5418 0.8903 0.9435
No log 2.4706 126 0.9931 0.5239 0.9931 0.9965
No log 2.5098 128 1.0006 0.5164 1.0006 1.0003
No log 2.5490 130 0.9428 0.5565 0.9428 0.9710
No log 2.5882 132 0.9686 0.4773 0.9686 0.9842
No log 2.6275 134 1.0382 0.4556 1.0382 1.0189
No log 2.6667 136 1.0113 0.4920 1.0113 1.0056
No log 2.7059 138 1.0617 0.5175 1.0617 1.0304
No log 2.7451 140 1.2730 0.4440 1.2730 1.1283
No log 2.7843 142 1.2842 0.4377 1.2842 1.1332
No log 2.8235 144 1.0403 0.5235 1.0403 1.0199
No log 2.8627 146 0.8743 0.5293 0.8743 0.9350
No log 2.9020 148 0.9584 0.4516 0.9584 0.9790
No log 2.9412 150 0.9957 0.4153 0.9957 0.9978
No log 2.9804 152 0.8888 0.4803 0.8888 0.9428
No log 3.0196 154 0.8859 0.5079 0.8859 0.9412
No log 3.0588 156 1.0052 0.5245 1.0052 1.0026
No log 3.0980 158 0.9953 0.5180 0.9953 0.9977
No log 3.1373 160 0.9499 0.5221 0.9499 0.9746
No log 3.1765 162 0.9760 0.4777 0.9760 0.9879
No log 3.2157 164 0.9746 0.4717 0.9746 0.9872
No log 3.2549 166 0.9259 0.5235 0.9259 0.9623
No log 3.2941 168 0.9789 0.5071 0.9789 0.9894
No log 3.3333 170 1.0175 0.5122 1.0175 1.0087
No log 3.3725 172 0.9951 0.4908 0.9951 0.9976
No log 3.4118 174 0.8876 0.4709 0.8876 0.9421
No log 3.4510 176 0.9356 0.4564 0.9356 0.9673
No log 3.4902 178 1.0225 0.4265 1.0225 1.0112
No log 3.5294 180 0.9906 0.4354 0.9906 0.9953
No log 3.5686 182 1.0064 0.4976 1.0064 1.0032
No log 3.6078 184 1.1146 0.4519 1.1146 1.0558
No log 3.6471 186 1.0741 0.4877 1.0741 1.0364
No log 3.6863 188 1.0130 0.4841 1.0130 1.0065
No log 3.7255 190 1.0344 0.4071 1.0344 1.0171
No log 3.7647 192 0.9908 0.4291 0.9908 0.9954
No log 3.8039 194 0.9572 0.4936 0.9572 0.9784
No log 3.8431 196 0.9575 0.4931 0.9575 0.9785
No log 3.8824 198 0.9550 0.4748 0.9550 0.9773
No log 3.9216 200 0.9387 0.5116 0.9387 0.9688
No log 3.9608 202 0.9394 0.4935 0.9394 0.9692
No log 4.0 204 0.9784 0.4645 0.9784 0.9891
No log 4.0392 206 0.9834 0.4534 0.9834 0.9916
No log 4.0784 208 0.9223 0.5117 0.9223 0.9604
No log 4.1176 210 0.9122 0.5101 0.9122 0.9551
No log 4.1569 212 0.8853 0.5110 0.8853 0.9409
No log 4.1961 214 0.8797 0.5123 0.8797 0.9379
No log 4.2353 216 0.8669 0.5048 0.8669 0.9311
No log 4.2745 218 0.9576 0.5263 0.9576 0.9786
No log 4.3137 220 1.0804 0.4696 1.0804 1.0394
No log 4.3529 222 1.1306 0.4543 1.1306 1.0633
No log 4.3922 224 1.0795 0.4602 1.0795 1.0390
No log 4.4314 226 0.9910 0.4657 0.9910 0.9955
No log 4.4706 228 0.9010 0.5005 0.9010 0.9492
No log 4.5098 230 0.8830 0.4872 0.8830 0.9397
No log 4.5490 232 0.9061 0.5314 0.9061 0.9519
No log 4.5882 234 0.9103 0.4993 0.9103 0.9541
No log 4.6275 236 0.9716 0.4950 0.9716 0.9857
No log 4.6667 238 1.0228 0.4615 1.0228 1.0113
No log 4.7059 240 1.0826 0.4538 1.0826 1.0405
No log 4.7451 242 1.1201 0.4635 1.1201 1.0583
No log 4.7843 244 1.1303 0.4648 1.1303 1.0631
No log 4.8235 246 1.1508 0.4279 1.1508 1.0728
No log 4.8627 248 1.0628 0.4508 1.0628 1.0309
No log 4.9020 250 0.9657 0.4995 0.9657 0.9827
No log 4.9412 252 0.8959 0.5136 0.8959 0.9465
No log 4.9804 254 0.8800 0.5008 0.8800 0.9381
No log 5.0196 256 0.8848 0.4964 0.8848 0.9406
No log 5.0588 258 0.8785 0.4837 0.8785 0.9373
No log 5.0980 260 0.9086 0.4730 0.9086 0.9532
No log 5.1373 262 0.9303 0.4837 0.9303 0.9645
No log 5.1765 264 0.9656 0.4581 0.9656 0.9827
No log 5.2157 266 1.0115 0.4361 1.0115 1.0057
No log 5.2549 268 1.0585 0.4415 1.0585 1.0288
No log 5.2941 270 1.0788 0.4212 1.0788 1.0387
No log 5.3333 272 1.0809 0.4242 1.0809 1.0397
No log 5.3725 274 1.0501 0.4521 1.0501 1.0247
No log 5.4118 276 0.9813 0.4081 0.9813 0.9906
No log 5.4510 278 0.9283 0.4299 0.9283 0.9635
No log 5.4902 280 0.8938 0.4282 0.8938 0.9454
No log 5.5294 282 0.8678 0.4492 0.8678 0.9316
No log 5.5686 284 0.8678 0.4355 0.8678 0.9316
No log 5.6078 286 0.9081 0.4378 0.9081 0.9529
No log 5.6471 288 1.0121 0.4165 1.0121 1.0060
No log 5.6863 290 1.0250 0.4057 1.0250 1.0124
No log 5.7255 292 0.9425 0.4378 0.9425 0.9708
No log 5.7647 294 0.9296 0.4920 0.9296 0.9642
No log 5.8039 296 0.9902 0.4838 0.9902 0.9951
No log 5.8431 298 1.0123 0.4768 1.0123 1.0061
No log 5.8824 300 1.0126 0.4773 1.0126 1.0063
No log 5.9216 302 0.9802 0.4723 0.9802 0.9900
No log 5.9608 304 0.9289 0.4713 0.9289 0.9638
No log 6.0 306 0.8907 0.4720 0.8907 0.9438
No log 6.0392 308 0.8875 0.4968 0.8875 0.9421
No log 6.0784 310 0.8795 0.5097 0.8795 0.9378
No log 6.1176 312 0.8910 0.4968 0.8910 0.9439
No log 6.1569 314 0.9162 0.4665 0.9162 0.9572
No log 6.1961 316 0.9334 0.4907 0.9334 0.9661
No log 6.2353 318 0.9608 0.4819 0.9608 0.9802
No log 6.2745 320 0.9441 0.4887 0.9441 0.9716
No log 6.3137 322 0.9319 0.4826 0.9319 0.9654
No log 6.3529 324 0.9151 0.4674 0.9151 0.9566
No log 6.3922 326 0.9004 0.4739 0.9004 0.9489
No log 6.4314 328 0.8825 0.4674 0.8825 0.9394
No log 6.4706 330 0.8694 0.4931 0.8694 0.9324
No log 6.5098 332 0.8525 0.4857 0.8525 0.9233
No log 6.5490 334 0.8571 0.4922 0.8571 0.9258
No log 6.5882 336 0.8984 0.4899 0.8984 0.9478
No log 6.6275 338 0.9820 0.4819 0.9820 0.9910
No log 6.6667 340 1.0828 0.4739 1.0828 1.0406
No log 6.7059 342 1.1232 0.4741 1.1232 1.0598
No log 6.7451 344 1.1055 0.4864 1.1055 1.0514
No log 6.7843 346 1.0656 0.4931 1.0656 1.0323
No log 6.8235 348 1.0132 0.495 1.0132 1.0066
No log 6.8627 350 0.9783 0.4676 0.9783 0.9891
No log 6.9020 352 0.9398 0.4676 0.9398 0.9694
No log 6.9412 354 0.9167 0.4676 0.9167 0.9574
No log 6.9804 356 0.9021 0.4788 0.9021 0.9498
No log 7.0196 358 0.9096 0.4765 0.9096 0.9537
No log 7.0588 360 0.9194 0.4594 0.9194 0.9588
No log 7.0980 362 0.9387 0.4463 0.9387 0.9689
No log 7.1373 364 0.9709 0.4903 0.9709 0.9854
No log 7.1765 366 0.9984 0.4575 0.9984 0.9992
No log 7.2157 368 0.9768 0.4593 0.9768 0.9883
No log 7.2549 370 0.9550 0.4587 0.9550 0.9773
No log 7.2941 372 0.9117 0.4820 0.9117 0.9548
No log 7.3333 374 0.8558 0.4701 0.8558 0.9251
No log 7.3725 376 0.8426 0.4783 0.8426 0.9179
No log 7.4118 378 0.8498 0.4783 0.8498 0.9218
No log 7.4510 380 0.8725 0.4666 0.8725 0.9341
No log 7.4902 382 0.9132 0.4847 0.9132 0.9556
No log 7.5294 384 0.9412 0.4758 0.9412 0.9701
No log 7.5686 386 0.9574 0.4753 0.9574 0.9785
No log 7.6078 388 0.9748 0.4708 0.9748 0.9873
No log 7.6471 390 1.0111 0.4691 1.0111 1.0055
No log 7.6863 392 1.0247 0.4372 1.0247 1.0123
No log 7.7255 394 1.0062 0.4717 1.0062 1.0031
No log 7.7647 396 0.9984 0.4685 0.9984 0.9992
No log 7.8039 398 0.9791 0.4575 0.9791 0.9895
No log 7.8431 400 0.9474 0.4584 0.9474 0.9734
No log 7.8824 402 0.9164 0.4695 0.9164 0.9573
No log 7.9216 404 0.8995 0.5053 0.8995 0.9484
No log 7.9608 406 0.8830 0.5297 0.8830 0.9397
No log 8.0 408 0.9004 0.4629 0.9004 0.9489
No log 8.0392 410 0.9397 0.4488 0.9397 0.9694
No log 8.0784 412 0.9670 0.4193 0.9670 0.9833
No log 8.1176 414 0.9581 0.4438 0.9581 0.9788
No log 8.1569 416 0.9369 0.4692 0.9369 0.9679
No log 8.1961 418 0.9115 0.5189 0.9115 0.9547
No log 8.2353 420 0.9083 0.4897 0.9083 0.9531
No log 8.2745 422 0.9156 0.4814 0.9156 0.9569
No log 8.3137 424 0.9196 0.4823 0.9196 0.9590
No log 8.3529 426 0.9234 0.4787 0.9234 0.9609
No log 8.3922 428 0.9428 0.4816 0.9428 0.9710
No log 8.4314 430 0.9431 0.4689 0.9431 0.9712
No log 8.4706 432 0.9364 0.4689 0.9364 0.9677
No log 8.5098 434 0.9260 0.4998 0.9260 0.9623
No log 8.5490 436 0.9258 0.4607 0.9258 0.9622
No log 8.5882 438 0.9335 0.4610 0.9335 0.9662
No log 8.6275 440 0.9448 0.4849 0.9448 0.9720
No log 8.6667 442 0.9525 0.4849 0.9525 0.9759
No log 8.7059 444 0.9478 0.4610 0.9478 0.9736
No log 8.7451 446 0.9401 0.4607 0.9401 0.9696
No log 8.7843 448 0.9352 0.4636 0.9352 0.9670
No log 8.8235 450 0.9342 0.4802 0.9342 0.9665
No log 8.8627 452 0.9268 0.4823 0.9268 0.9627
No log 8.9020 454 0.9231 0.4809 0.9231 0.9608
No log 8.9412 456 0.9204 0.4925 0.9204 0.9594
No log 8.9804 458 0.9080 0.4925 0.9080 0.9529
No log 9.0196 460 0.8887 0.4822 0.8887 0.9427
No log 9.0588 462 0.8685 0.4728 0.8685 0.9319
No log 9.0980 464 0.8500 0.4973 0.8500 0.9219
No log 9.1373 466 0.8348 0.5123 0.8348 0.9137
No log 9.1765 468 0.8306 0.5123 0.8306 0.9114
No log 9.2157 470 0.8326 0.5095 0.8326 0.9125
No log 9.2549 472 0.8372 0.5295 0.8372 0.9150
No log 9.2941 474 0.8422 0.4966 0.8422 0.9177
No log 9.3333 476 0.8475 0.5054 0.8475 0.9206
No log 9.3725 478 0.8544 0.4973 0.8544 0.9243
No log 9.4118 480 0.8642 0.4795 0.8642 0.9296
No log 9.4510 482 0.8718 0.4666 0.8718 0.9337
No log 9.4902 484 0.8780 0.4666 0.8780 0.9370
No log 9.5294 486 0.8840 0.4666 0.8840 0.9402
No log 9.5686 488 0.8912 0.4666 0.8912 0.9440
No log 9.6078 490 0.8934 0.4907 0.8934 0.9452
No log 9.6471 492 0.8922 0.4913 0.8922 0.9446
No log 9.6863 494 0.8925 0.4934 0.8925 0.9447
No log 9.7255 496 0.8935 0.4942 0.8935 0.9452
No log 9.7647 498 0.8951 0.5072 0.8951 0.9461
0.399 9.8039 500 0.8955 0.5245 0.8955 0.9463
0.399 9.8431 502 0.8958 0.5255 0.8958 0.9465
0.399 9.8824 504 0.8954 0.5255 0.8954 0.9463
0.399 9.9216 506 0.8953 0.5255 0.8953 0.9462
0.399 9.9608 508 0.8954 0.5255 0.8954 0.9463
0.399 10.0 510 0.8955 0.5255 0.8955 0.9463

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run1_AugV5_k9_task2_organization

Finetuned
(4023)
this model