ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k8_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9451
  • Qwk: 0.4576
  • Mse: 0.9451
  • Rmse: 0.9722

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0769 2 4.8459 0.0064 4.8459 2.2013
No log 0.1538 4 3.1151 -0.0625 3.1151 1.7650
No log 0.2308 6 1.7017 0.0504 1.7017 1.3045
No log 0.3077 8 1.5782 -0.0508 1.5782 1.2563
No log 0.3846 10 1.2904 0.0666 1.2904 1.1360
No log 0.4615 12 1.3527 -0.0757 1.3527 1.1630
No log 0.5385 14 1.6222 0.0 1.6222 1.2736
No log 0.6154 16 1.5780 0.0 1.5780 1.2562
No log 0.6923 18 1.3871 -0.0149 1.3871 1.1778
No log 0.7692 20 1.2612 0.1753 1.2612 1.1230
No log 0.8462 22 1.2350 0.1397 1.2350 1.1113
No log 0.9231 24 1.3712 0.0561 1.3712 1.1710
No log 1.0 26 1.3531 0.0342 1.3531 1.1632
No log 1.0769 28 1.2235 0.2886 1.2235 1.1061
No log 1.1538 30 1.1670 0.2784 1.1670 1.0803
No log 1.2308 32 1.1681 0.1458 1.1681 1.0808
No log 1.3077 34 1.1690 0.2584 1.1690 1.0812
No log 1.3846 36 1.1727 0.1638 1.1727 1.0829
No log 1.4615 38 1.2101 0.1865 1.2101 1.1000
No log 1.5385 40 1.2069 0.1735 1.2069 1.0986
No log 1.6154 42 1.1310 0.1809 1.1310 1.0635
No log 1.6923 44 1.0691 0.3421 1.0691 1.0340
No log 1.7692 46 1.1262 0.2364 1.1262 1.0612
No log 1.8462 48 1.1588 0.1458 1.1588 1.0765
No log 1.9231 50 1.3259 0.1260 1.3259 1.1515
No log 2.0 52 1.6163 0.0800 1.6163 1.2714
No log 2.0769 54 1.5475 0.1198 1.5475 1.2440
No log 2.1538 56 1.2395 0.1562 1.2395 1.1133
No log 2.2308 58 1.1374 0.2372 1.1374 1.0665
No log 2.3077 60 1.0502 0.2134 1.0502 1.0248
No log 2.3846 62 1.0714 0.3115 1.0714 1.0351
No log 2.4615 64 1.0774 0.4405 1.0774 1.0380
No log 2.5385 66 1.0478 0.4311 1.0478 1.0236
No log 2.6154 68 1.1091 0.2285 1.1091 1.0531
No log 2.6923 70 1.1674 0.1959 1.1674 1.0804
No log 2.7692 72 1.1498 0.2098 1.1498 1.0723
No log 2.8462 74 1.0350 0.4434 1.0350 1.0173
No log 2.9231 76 0.9890 0.4548 0.9890 0.9945
No log 3.0 78 0.9957 0.3543 0.9957 0.9978
No log 3.0769 80 1.0168 0.4872 1.0168 1.0084
No log 3.1538 82 1.0254 0.4589 1.0254 1.0126
No log 3.2308 84 0.9455 0.4711 0.9455 0.9723
No log 3.3077 86 0.9279 0.4667 0.9279 0.9633
No log 3.3846 88 0.9662 0.4045 0.9662 0.9830
No log 3.4615 90 0.9476 0.3956 0.9476 0.9735
No log 3.5385 92 0.9284 0.4656 0.9284 0.9636
No log 3.6154 94 1.1353 0.4692 1.1353 1.0655
No log 3.6923 96 1.3347 0.4284 1.3347 1.1553
No log 3.7692 98 1.1981 0.4565 1.1981 1.0946
No log 3.8462 100 0.9231 0.4714 0.9231 0.9608
No log 3.9231 102 0.9209 0.4941 0.9209 0.9597
No log 4.0 104 0.9215 0.4273 0.9215 0.9600
No log 4.0769 106 0.8645 0.3943 0.8645 0.9298
No log 4.1538 108 1.0034 0.4057 1.0034 1.0017
No log 4.2308 110 1.0829 0.4032 1.0829 1.0406
No log 4.3077 112 1.0129 0.4632 1.0129 1.0064
No log 4.3846 114 0.8756 0.4465 0.8756 0.9357
No log 4.4615 116 0.8090 0.4321 0.8090 0.8994
No log 4.5385 118 0.8092 0.5107 0.8092 0.8995
No log 4.6154 120 0.8269 0.4808 0.8269 0.9094
No log 4.6923 122 0.9227 0.4790 0.9227 0.9606
No log 4.7692 124 0.9785 0.4668 0.9785 0.9892
No log 4.8462 126 0.8627 0.5094 0.8627 0.9288
No log 4.9231 128 0.8612 0.5263 0.8612 0.9280
No log 5.0 130 1.0890 0.4820 1.0890 1.0435
No log 5.0769 132 1.0301 0.4085 1.0301 1.0149
No log 5.1538 134 0.8508 0.4624 0.8508 0.9224
No log 5.2308 136 0.8463 0.4572 0.8463 0.9200
No log 5.3077 138 1.1195 0.4468 1.1195 1.0581
No log 5.3846 140 1.1389 0.4080 1.1389 1.0672
No log 5.4615 142 0.9934 0.4989 0.9934 0.9967
No log 5.5385 144 0.9016 0.4615 0.9016 0.9495
No log 5.6154 146 0.8690 0.5235 0.8690 0.9322
No log 5.6923 148 0.8557 0.4327 0.8557 0.9250
No log 5.7692 150 0.8858 0.4945 0.8858 0.9412
No log 5.8462 152 0.9637 0.5166 0.9637 0.9817
No log 5.9231 154 0.9436 0.4646 0.9436 0.9714
No log 6.0 156 0.8882 0.3908 0.8882 0.9424
No log 6.0769 158 0.8944 0.4534 0.8944 0.9457
No log 6.1538 160 0.9339 0.4308 0.9339 0.9664
No log 6.2308 162 0.9274 0.4563 0.9274 0.9630
No log 6.3077 164 0.8299 0.4385 0.8299 0.9110
No log 6.3846 166 0.9458 0.4790 0.9458 0.9725
No log 6.4615 168 1.1202 0.4784 1.1202 1.0584
No log 6.5385 170 1.0964 0.5066 1.0964 1.0471
No log 6.6154 172 0.9839 0.4435 0.9839 0.9919
No log 6.6923 174 0.8454 0.4575 0.8454 0.9194
No log 6.7692 176 0.8137 0.5393 0.8137 0.9020
No log 6.8462 178 0.8054 0.5399 0.8054 0.8975
No log 6.9231 180 0.8078 0.3935 0.8078 0.8988
No log 7.0 182 0.9439 0.5253 0.9439 0.9715
No log 7.0769 184 1.0224 0.5293 1.0224 1.0112
No log 7.1538 186 0.9585 0.5301 0.9585 0.9790
No log 7.2308 188 0.8855 0.4920 0.8855 0.9410
No log 7.3077 190 0.8332 0.3852 0.8332 0.9128
No log 7.3846 192 0.8177 0.4086 0.8177 0.9043
No log 7.4615 194 0.8245 0.3811 0.8245 0.9080
No log 7.5385 196 0.8743 0.3880 0.8743 0.9350
No log 7.6154 198 0.9313 0.5098 0.9313 0.9650
No log 7.6923 200 0.8926 0.5014 0.8926 0.9448
No log 7.7692 202 0.8269 0.4763 0.8269 0.9093
No log 7.8462 204 0.7917 0.5351 0.7917 0.8898
No log 7.9231 206 0.8871 0.5614 0.8871 0.9418
No log 8.0 208 1.1694 0.5491 1.1694 1.0814
No log 8.0769 210 1.2706 0.5082 1.2706 1.1272
No log 8.1538 212 1.0750 0.5341 1.0750 1.0368
No log 8.2308 214 0.8655 0.5738 0.8655 0.9303
No log 8.3077 216 0.8147 0.4726 0.8147 0.9026
No log 8.3846 218 0.8277 0.4144 0.8277 0.9098
No log 8.4615 220 0.8802 0.4006 0.8802 0.9382
No log 8.5385 222 0.9052 0.4202 0.9052 0.9514
No log 8.6154 224 0.9335 0.4924 0.9335 0.9662
No log 8.6923 226 0.8718 0.4754 0.8718 0.9337
No log 8.7692 228 0.8347 0.3685 0.8347 0.9136
No log 8.8462 230 0.9309 0.3806 0.9309 0.9649
No log 8.9231 232 0.9113 0.4235 0.9113 0.9546
No log 9.0 234 0.8447 0.3629 0.8447 0.9191
No log 9.0769 236 0.8354 0.3700 0.8354 0.9140
No log 9.1538 238 0.8565 0.4640 0.8565 0.9255
No log 9.2308 240 0.8313 0.3957 0.8313 0.9118
No log 9.3077 242 0.7987 0.4526 0.7987 0.8937
No log 9.3846 244 0.7963 0.4519 0.7963 0.8923
No log 9.4615 246 0.7926 0.4142 0.7926 0.8903
No log 9.5385 248 0.8095 0.4242 0.8095 0.8997
No log 9.6154 250 0.8261 0.4470 0.8261 0.9089
No log 9.6923 252 0.8121 0.4493 0.8121 0.9012
No log 9.7692 254 0.8333 0.4698 0.8333 0.9129
No log 9.8462 256 0.8521 0.4698 0.8521 0.9231
No log 9.9231 258 0.8516 0.4308 0.8516 0.9228
No log 10.0 260 0.8320 0.4124 0.8320 0.9122
No log 10.0769 262 0.8656 0.4835 0.8656 0.9304
No log 10.1538 264 0.9032 0.5130 0.9032 0.9504
No log 10.2308 266 0.8454 0.5025 0.8454 0.9195
No log 10.3077 268 0.7992 0.5924 0.7992 0.8940
No log 10.3846 270 0.8008 0.5181 0.8008 0.8949
No log 10.4615 272 0.8056 0.5495 0.8056 0.8975
No log 10.5385 274 0.8030 0.4690 0.8030 0.8961
No log 10.6154 276 0.8043 0.4996 0.8043 0.8969
No log 10.6923 278 0.8569 0.4507 0.8569 0.9257
No log 10.7692 280 0.8679 0.4572 0.8679 0.9316
No log 10.8462 282 0.9047 0.4783 0.9047 0.9512
No log 10.9231 284 0.8770 0.4165 0.8770 0.9365
No log 11.0 286 0.8632 0.3957 0.8632 0.9291
No log 11.0769 288 0.9083 0.3551 0.9083 0.9531
No log 11.1538 290 0.9376 0.3335 0.9376 0.9683
No log 11.2308 292 0.9290 0.3354 0.9290 0.9639
No log 11.3077 294 0.9151 0.3074 0.9151 0.9566
No log 11.3846 296 0.8838 0.2834 0.8838 0.9401
No log 11.4615 298 0.8862 0.2834 0.8862 0.9414
No log 11.5385 300 0.8769 0.3174 0.8769 0.9364
No log 11.6154 302 0.8826 0.3957 0.8826 0.9395
No log 11.6923 304 0.9025 0.3777 0.9025 0.9500
No log 11.7692 306 0.9266 0.3377 0.9266 0.9626
No log 11.8462 308 0.9012 0.3474 0.9012 0.9493
No log 11.9231 310 0.8935 0.3372 0.8935 0.9453
No log 12.0 312 0.8486 0.3854 0.8486 0.9212
No log 12.0769 314 0.8238 0.3854 0.8238 0.9076
No log 12.1538 316 0.7945 0.4359 0.7945 0.8913
No log 12.2308 318 0.7982 0.5279 0.7982 0.8934
No log 12.3077 320 0.7934 0.4691 0.7934 0.8907
No log 12.3846 322 0.8345 0.4337 0.8345 0.9135
No log 12.4615 324 0.9168 0.4614 0.9168 0.9575
No log 12.5385 326 0.9219 0.4394 0.9219 0.9602
No log 12.6154 328 0.8496 0.3278 0.8496 0.9217
No log 12.6923 330 0.8253 0.3570 0.8253 0.9084
No log 12.7692 332 0.8192 0.3570 0.8192 0.9051
No log 12.8462 334 0.8264 0.3689 0.8264 0.9091
No log 12.9231 336 0.8419 0.4666 0.8419 0.9175
No log 13.0 338 0.8333 0.5073 0.8333 0.9128
No log 13.0769 340 0.7873 0.5073 0.7873 0.8873
No log 13.1538 342 0.7345 0.5579 0.7345 0.8571
No log 13.2308 344 0.7385 0.5443 0.7385 0.8593
No log 13.3077 346 0.7502 0.5186 0.7502 0.8661
No log 13.3846 348 0.7924 0.5386 0.7924 0.8902
No log 13.4615 350 0.8314 0.5267 0.8314 0.9118
No log 13.5385 352 0.8338 0.5076 0.8338 0.9131
No log 13.6154 354 0.8067 0.3965 0.8067 0.8982
No log 13.6923 356 0.8166 0.4400 0.8166 0.9036
No log 13.7692 358 0.8172 0.4368 0.8172 0.9040
No log 13.8462 360 0.8146 0.4804 0.8146 0.9026
No log 13.9231 362 0.8042 0.5028 0.8042 0.8968
No log 14.0 364 0.7786 0.4608 0.7786 0.8824
No log 14.0769 366 0.7683 0.4839 0.7683 0.8765
No log 14.1538 368 0.7784 0.5552 0.7784 0.8822
No log 14.2308 370 0.8073 0.4889 0.8073 0.8985
No log 14.3077 372 0.8286 0.4910 0.8286 0.9103
No log 14.3846 374 0.8379 0.5083 0.8379 0.9154
No log 14.4615 376 0.8462 0.4773 0.8462 0.9199
No log 14.5385 378 0.9117 0.5071 0.9117 0.9548
No log 14.6154 380 0.8925 0.4703 0.8925 0.9447
No log 14.6923 382 0.8670 0.4201 0.8670 0.9312
No log 14.7692 384 0.8533 0.4202 0.8533 0.9238
No log 14.8462 386 0.8587 0.3729 0.8587 0.9267
No log 14.9231 388 0.8565 0.3478 0.8565 0.9254
No log 15.0 390 0.8253 0.3747 0.8253 0.9084
No log 15.0769 392 0.8152 0.4548 0.8152 0.9029
No log 15.1538 394 0.8277 0.4254 0.8277 0.9098
No log 15.2308 396 0.8600 0.4053 0.8600 0.9274
No log 15.3077 398 0.8954 0.3786 0.8954 0.9463
No log 15.3846 400 0.9037 0.3993 0.9037 0.9507
No log 15.4615 402 0.9093 0.3993 0.9093 0.9536
No log 15.5385 404 0.8822 0.4127 0.8822 0.9393
No log 15.6154 406 0.8355 0.4811 0.8355 0.9141
No log 15.6923 408 0.8039 0.4946 0.8039 0.8966
No log 15.7692 410 0.7805 0.4482 0.7805 0.8834
No log 15.8462 412 0.7978 0.4743 0.7978 0.8932
No log 15.9231 414 0.8798 0.4781 0.8798 0.9380
No log 16.0 416 0.8925 0.4663 0.8925 0.9447
No log 16.0769 418 0.8123 0.4822 0.8123 0.9013
No log 16.1538 420 0.7830 0.5424 0.7830 0.8849
No log 16.2308 422 0.7529 0.5403 0.7529 0.8677
No log 16.3077 424 0.7527 0.5479 0.7527 0.8676
No log 16.3846 426 0.7723 0.4811 0.7723 0.8788
No log 16.4615 428 0.8077 0.4302 0.8077 0.8987
No log 16.5385 430 0.8148 0.4302 0.8148 0.9026
No log 16.6154 432 0.8247 0.4302 0.8247 0.9081
No log 16.6923 434 0.7950 0.4931 0.7950 0.8916
No log 16.7692 436 0.7621 0.5213 0.7621 0.8730
No log 16.8462 438 0.7571 0.5213 0.7571 0.8701
No log 16.9231 440 0.7762 0.4920 0.7762 0.8810
No log 17.0 442 0.8354 0.5147 0.8354 0.9140
No log 17.0769 444 0.8604 0.4375 0.8604 0.9276
No log 17.1538 446 0.8151 0.4732 0.8151 0.9028
No log 17.2308 448 0.7574 0.4243 0.7574 0.8703
No log 17.3077 450 0.7487 0.4826 0.7487 0.8653
No log 17.3846 452 0.7461 0.4826 0.7461 0.8638
No log 17.4615 454 0.7489 0.5240 0.7489 0.8654
No log 17.5385 456 0.8760 0.4663 0.8760 0.9359
No log 17.6154 458 0.9547 0.5098 0.9547 0.9771
No log 17.6923 460 0.8928 0.4834 0.8928 0.9449
No log 17.7692 462 0.7758 0.4754 0.7758 0.8808
No log 17.8462 464 0.7166 0.5649 0.7166 0.8465
No log 17.9231 466 0.7135 0.4993 0.7135 0.8447
No log 18.0 468 0.7224 0.5242 0.7224 0.8499
No log 18.0769 470 0.7522 0.5295 0.7522 0.8673
No log 18.1538 472 0.7908 0.4774 0.7908 0.8893
No log 18.2308 474 0.7783 0.5365 0.7783 0.8822
No log 18.3077 476 0.7639 0.4746 0.7639 0.8740
No log 18.3846 478 0.7722 0.4931 0.7721 0.8787
No log 18.4615 480 0.7906 0.4568 0.7906 0.8891
No log 18.5385 482 0.8157 0.4302 0.8157 0.9032
No log 18.6154 484 0.8331 0.4631 0.8331 0.9128
No log 18.6923 486 0.8126 0.4754 0.8126 0.9014
No log 18.7692 488 0.7770 0.5346 0.7770 0.8815
No log 18.8462 490 0.7876 0.5324 0.7876 0.8875
No log 18.9231 492 0.7630 0.5324 0.7630 0.8735
No log 19.0 494 0.7423 0.5663 0.7423 0.8616
No log 19.0769 496 0.7594 0.5283 0.7594 0.8715
No log 19.1538 498 0.7928 0.5447 0.7928 0.8904
0.3447 19.2308 500 0.7959 0.5041 0.7959 0.8922
0.3447 19.3077 502 0.7956 0.5086 0.7956 0.8919
0.3447 19.3846 504 0.8228 0.4774 0.8228 0.9071
0.3447 19.4615 506 0.8301 0.4774 0.8301 0.9111
0.3447 19.5385 508 0.8486 0.4774 0.8486 0.9212
0.3447 19.6154 510 0.9081 0.4521 0.9081 0.9529
0.3447 19.6923 512 0.9451 0.4576 0.9451 0.9722

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k8_task2_organization

Finetuned
(4023)
this model