ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k8_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8395
  • Qwk: 0.3983
  • Mse: 0.8395
  • Rmse: 0.9162

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0769 2 4.6183 0.0010 4.6183 2.1490
No log 0.1538 4 2.6915 0.0088 2.6915 1.6406
No log 0.2308 6 1.6409 0.0372 1.6409 1.2810
No log 0.3077 8 1.3863 -0.0809 1.3863 1.1774
No log 0.3846 10 1.3433 0.0346 1.3433 1.1590
No log 0.4615 12 1.3203 0.0639 1.3203 1.1490
No log 0.5385 14 1.6131 0.0 1.6131 1.2701
No log 0.6154 16 2.2292 -0.0380 2.2292 1.4930
No log 0.6923 18 2.0569 0.0 2.0569 1.4342
No log 0.7692 20 1.9918 0.0372 1.9918 1.4113
No log 0.8462 22 1.9310 0.0372 1.9310 1.3896
No log 0.9231 24 1.6950 0.0 1.6950 1.3019
No log 1.0 26 1.5565 0.0 1.5565 1.2476
No log 1.0769 28 1.3891 -0.0149 1.3891 1.1786
No log 1.1538 30 1.4566 -0.0149 1.4566 1.2069
No log 1.2308 32 1.5147 0.0585 1.5147 1.2307
No log 1.3077 34 1.6367 0.1060 1.6367 1.2793
No log 1.3846 36 1.3206 0.0393 1.3206 1.1492
No log 1.4615 38 1.1342 0.3183 1.1342 1.0650
No log 1.5385 40 1.1373 0.2579 1.1373 1.0665
No log 1.6154 42 1.2459 0.0537 1.2459 1.1162
No log 1.6923 44 1.2742 0.0057 1.2742 1.1288
No log 1.7692 46 1.1179 0.2886 1.1179 1.0573
No log 1.8462 48 1.0140 0.3100 1.0140 1.0070
No log 1.9231 50 1.0141 0.3663 1.0141 1.0070
No log 2.0 52 0.9968 0.3663 0.9968 0.9984
No log 2.0769 54 0.9833 0.3663 0.9833 0.9916
No log 2.1538 56 0.9788 0.3859 0.9788 0.9893
No log 2.2308 58 0.9210 0.3307 0.9210 0.9597
No log 2.3077 60 0.9337 0.4222 0.9337 0.9663
No log 2.3846 62 0.9333 0.4222 0.9333 0.9661
No log 2.4615 64 1.2117 0.1928 1.2117 1.1008
No log 2.5385 66 1.2342 0.1928 1.2342 1.1109
No log 2.6154 68 0.9444 0.4764 0.9444 0.9718
No log 2.6923 70 0.8483 0.3629 0.8483 0.9210
No log 2.7692 72 0.9602 0.2658 0.9602 0.9799
No log 2.8462 74 0.9444 0.2640 0.9444 0.9718
No log 2.9231 76 0.8800 0.3921 0.8800 0.9381
No log 3.0 78 0.9439 0.4012 0.9439 0.9715
No log 3.0769 80 0.9766 0.4819 0.9766 0.9882
No log 3.1538 82 0.8943 0.4645 0.8943 0.9457
No log 3.2308 84 0.8558 0.4461 0.8558 0.9251
No log 3.3077 86 0.8591 0.4359 0.8591 0.9269
No log 3.3846 88 0.8563 0.3979 0.8563 0.9254
No log 3.4615 90 0.7955 0.4316 0.7955 0.8919
No log 3.5385 92 0.8069 0.4401 0.8069 0.8983
No log 3.6154 94 0.8860 0.5156 0.8860 0.9413
No log 3.6923 96 0.9130 0.5279 0.9130 0.9555
No log 3.7692 98 0.8529 0.3974 0.8529 0.9236
No log 3.8462 100 1.1062 0.3384 1.1062 1.0517
No log 3.9231 102 1.3142 0.1080 1.3142 1.1464
No log 4.0 104 1.1620 0.1748 1.1620 1.0780
No log 4.0769 106 0.8057 0.4197 0.8057 0.8976
No log 4.1538 108 0.8967 0.5529 0.8967 0.9469
No log 4.2308 110 0.8838 0.5173 0.8838 0.9401
No log 4.3077 112 0.7904 0.4652 0.7904 0.8891
No log 4.3846 114 0.8651 0.4944 0.8651 0.9301
No log 4.4615 116 0.8261 0.5086 0.8261 0.9089
No log 4.5385 118 0.9370 0.4672 0.9370 0.9680
No log 4.6154 120 1.3524 0.4086 1.3524 1.1629
No log 4.6923 122 1.4519 0.3172 1.4519 1.2049
No log 4.7692 124 1.2421 0.4077 1.2421 1.1145
No log 4.8462 126 0.9421 0.4323 0.9421 0.9706
No log 4.9231 128 0.8703 0.4749 0.8703 0.9329
No log 5.0 130 0.8839 0.4482 0.8839 0.9402
No log 5.0769 132 0.9676 0.5173 0.9676 0.9837
No log 5.1538 134 1.0313 0.5156 1.0313 1.0155
No log 5.2308 136 1.0060 0.5354 1.0060 1.0030
No log 5.3077 138 0.8308 0.4894 0.8308 0.9115
No log 5.3846 140 0.8735 0.5283 0.8735 0.9346
No log 5.4615 142 1.0642 0.4387 1.0642 1.0316
No log 5.5385 144 0.9434 0.4676 0.9434 0.9713
No log 5.6154 146 0.7837 0.4847 0.7837 0.8853
No log 5.6923 148 0.8692 0.4858 0.8692 0.9323
No log 5.7692 150 1.0888 0.5563 1.0888 1.0435
No log 5.8462 152 1.3006 0.4302 1.3006 1.1404
No log 5.9231 154 1.1541 0.5505 1.1541 1.0743
No log 6.0 156 0.8706 0.4482 0.8706 0.9330
No log 6.0769 158 0.8707 0.4471 0.8707 0.9331
No log 6.1538 160 0.9463 0.4676 0.9463 0.9728
No log 6.2308 162 0.8804 0.4471 0.8804 0.9383
No log 6.3077 164 0.8934 0.3525 0.8934 0.9452
No log 6.3846 166 0.9451 0.4102 0.9451 0.9722
No log 6.4615 168 0.9315 0.3508 0.9315 0.9652
No log 6.5385 170 0.8961 0.2944 0.8961 0.9466
No log 6.6154 172 0.8864 0.3960 0.8864 0.9415
No log 6.6923 174 0.8740 0.3200 0.8740 0.9349
No log 6.7692 176 0.8903 0.4311 0.8903 0.9435
No log 6.8462 178 0.9667 0.5029 0.9667 0.9832
No log 6.9231 180 0.9486 0.4398 0.9486 0.9740
No log 7.0 182 0.8840 0.4343 0.8840 0.9402
No log 7.0769 184 0.8452 0.3483 0.8452 0.9193
No log 7.1538 186 0.8534 0.4096 0.8534 0.9238
No log 7.2308 188 0.8322 0.4079 0.8322 0.9123
No log 7.3077 190 0.9347 0.4865 0.9347 0.9668
No log 7.3846 192 1.0289 0.5030 1.0289 1.0144
No log 7.4615 194 0.9706 0.4772 0.9706 0.9852
No log 7.5385 196 0.8598 0.4700 0.8598 0.9272
No log 7.6154 198 0.8068 0.4813 0.8068 0.8982
No log 7.6923 200 0.8546 0.4078 0.8546 0.9244
No log 7.7692 202 0.8643 0.4379 0.8643 0.9297
No log 7.8462 204 0.7887 0.5025 0.7887 0.8881
No log 7.9231 206 0.8229 0.4796 0.8229 0.9071
No log 8.0 208 0.9449 0.5 0.9449 0.9721
No log 8.0769 210 0.9933 0.4822 0.9933 0.9966
No log 8.1538 212 0.8974 0.4334 0.8974 0.9473
No log 8.2308 214 0.8220 0.4282 0.8220 0.9066
No log 8.3077 216 0.8234 0.4077 0.8234 0.9074
No log 8.3846 218 0.8125 0.4077 0.8125 0.9014
No log 8.4615 220 0.8386 0.4700 0.8386 0.9158
No log 8.5385 222 1.0036 0.4781 1.0036 1.0018
No log 8.6154 224 1.0321 0.5334 1.0321 1.0159
No log 8.6923 226 0.8975 0.5028 0.8975 0.9474
No log 8.7692 228 0.7461 0.5089 0.7461 0.8638
No log 8.8462 230 0.7383 0.4794 0.7383 0.8592
No log 8.9231 232 0.7424 0.4926 0.7424 0.8616
No log 9.0 234 0.7326 0.4945 0.7326 0.8559
No log 9.0769 236 0.7673 0.4948 0.7673 0.8760
No log 9.1538 238 0.8661 0.4631 0.8661 0.9306
No log 9.2308 240 0.8426 0.4560 0.8426 0.9179
No log 9.3077 242 0.7977 0.4519 0.7977 0.8931
No log 9.3846 244 0.8087 0.4363 0.8087 0.8993
No log 9.4615 246 0.8493 0.5093 0.8493 0.9216
No log 9.5385 248 0.8317 0.5437 0.8317 0.9119
No log 9.6154 250 0.7958 0.4219 0.7958 0.8921
No log 9.6923 252 0.8151 0.4916 0.8151 0.9028
No log 9.7692 254 0.8370 0.4681 0.8370 0.9149
No log 9.8462 256 0.8202 0.4916 0.8202 0.9057
No log 9.9231 258 0.8241 0.4241 0.8241 0.9078
No log 10.0 260 0.8408 0.4241 0.8408 0.9170
No log 10.0769 262 0.8865 0.3687 0.8865 0.9415
No log 10.1538 264 0.9354 0.3985 0.9354 0.9672
No log 10.2308 266 0.9941 0.3691 0.9941 0.9970
No log 10.3077 268 1.0000 0.3691 1.0000 1.0000
No log 10.3846 270 0.9574 0.4262 0.9574 0.9785
No log 10.4615 272 0.8841 0.4142 0.8841 0.9402
No log 10.5385 274 0.8663 0.3821 0.8663 0.9307
No log 10.6154 276 0.9381 0.4926 0.9381 0.9685
No log 10.6923 278 0.8981 0.4893 0.8981 0.9477
No log 10.7692 280 0.8051 0.4926 0.8051 0.8973
No log 10.8462 282 0.7904 0.4534 0.7904 0.8890
No log 10.9231 284 0.8202 0.4612 0.8202 0.9056
No log 11.0 286 0.8464 0.4704 0.8464 0.9200
No log 11.0769 288 0.8595 0.4366 0.8595 0.9271
No log 11.1538 290 0.8562 0.4275 0.8562 0.9253
No log 11.2308 292 0.8667 0.4045 0.8667 0.9310
No log 11.3077 294 0.8714 0.4045 0.8714 0.9335
No log 11.3846 296 0.8588 0.4813 0.8588 0.9267
No log 11.4615 298 0.8516 0.4945 0.8516 0.9228
No log 11.5385 300 0.8211 0.3948 0.8211 0.9061
No log 11.6154 302 0.8414 0.4796 0.8414 0.9173
No log 11.6923 304 0.8425 0.4796 0.8425 0.9179
No log 11.7692 306 0.8734 0.4954 0.8734 0.9346
No log 11.8462 308 0.8423 0.4595 0.8423 0.9178
No log 11.9231 310 0.8088 0.3852 0.8088 0.8993
No log 12.0 312 0.8304 0.4098 0.8304 0.9113
No log 12.0769 314 0.9175 0.4741 0.9175 0.9578
No log 12.1538 316 0.9361 0.4717 0.9361 0.9675
No log 12.2308 318 0.8919 0.3627 0.8919 0.9444
No log 12.3077 320 0.9926 0.4259 0.9926 0.9963
No log 12.3846 322 1.1236 0.4552 1.1236 1.0600
No log 12.4615 324 1.1239 0.4311 1.1239 1.0601
No log 12.5385 326 1.0108 0.3560 1.0108 1.0054
No log 12.6154 328 0.9169 0.3596 0.9169 0.9576
No log 12.6923 330 0.9123 0.3283 0.9123 0.9551
No log 12.7692 332 0.9001 0.3283 0.9001 0.9487
No log 12.8462 334 0.8838 0.3845 0.8838 0.9401
No log 12.9231 336 0.9431 0.4069 0.9431 0.9711
No log 13.0 338 0.9584 0.4295 0.9584 0.9790
No log 13.0769 340 0.9103 0.4203 0.9103 0.9541
No log 13.1538 342 0.8542 0.4512 0.8542 0.9242
No log 13.2308 344 0.8456 0.3885 0.8456 0.9196
No log 13.3077 346 0.8387 0.4023 0.8387 0.9158
No log 13.3846 348 0.8590 0.3723 0.8590 0.9268
No log 13.4615 350 0.8671 0.3681 0.8671 0.9312
No log 13.5385 352 0.8725 0.4122 0.8725 0.9341
No log 13.6154 354 0.9228 0.4165 0.9228 0.9606
No log 13.6923 356 0.9279 0.4028 0.9279 0.9633
No log 13.7692 358 0.8923 0.4418 0.8923 0.9446
No log 13.8462 360 0.8874 0.3627 0.8874 0.9420
No log 13.9231 362 0.8845 0.3771 0.8845 0.9405
No log 14.0 364 0.8848 0.3671 0.8848 0.9406
No log 14.0769 366 0.8711 0.3747 0.8711 0.9333
No log 14.1538 368 0.8746 0.4476 0.8746 0.9352
No log 14.2308 370 0.8714 0.4931 0.8714 0.9335
No log 14.3077 372 0.8328 0.4931 0.8328 0.9126
No log 14.3846 374 0.8066 0.3852 0.8066 0.8981
No log 14.4615 376 0.8101 0.4158 0.8101 0.9001
No log 14.5385 378 0.8059 0.3983 0.8059 0.8977
No log 14.6154 380 0.8317 0.4340 0.8317 0.9120
No log 14.6923 382 0.8588 0.4340 0.8588 0.9267
No log 14.7692 384 0.8721 0.4340 0.8721 0.9338
No log 14.8462 386 0.8424 0.4340 0.8424 0.9178
No log 14.9231 388 0.8113 0.4142 0.8113 0.9007
No log 15.0 390 0.8364 0.4292 0.8364 0.9145
No log 15.0769 392 0.8602 0.4686 0.8602 0.9274
No log 15.1538 394 0.8561 0.4356 0.8561 0.9253
No log 15.2308 396 0.8543 0.4241 0.8543 0.9243
No log 15.3077 398 0.8989 0.4470 0.8989 0.9481
No log 15.3846 400 0.9577 0.4833 0.9577 0.9786
No log 15.4615 402 0.9406 0.5264 0.9406 0.9699
No log 15.5385 404 0.8697 0.3724 0.8697 0.9326
No log 15.6154 406 0.8397 0.4945 0.8397 0.9164
No log 15.6923 408 0.8394 0.4960 0.8394 0.9162
No log 15.7692 410 0.8257 0.4945 0.8257 0.9087
No log 15.8462 412 0.8310 0.4045 0.8310 0.9116
No log 15.9231 414 0.8498 0.4369 0.8498 0.9218
No log 16.0 416 0.8493 0.4366 0.8493 0.9216
No log 16.0769 418 0.8483 0.4465 0.8483 0.9210
No log 16.1538 420 0.8483 0.4568 0.8483 0.9210
No log 16.2308 422 0.8320 0.4470 0.8320 0.9122
No log 16.3077 424 0.8235 0.4470 0.8235 0.9075
No log 16.3846 426 0.8257 0.4470 0.8257 0.9087
No log 16.4615 428 0.8345 0.4681 0.8345 0.9135
No log 16.5385 430 0.8365 0.4278 0.8365 0.9146
No log 16.6154 432 0.8473 0.3705 0.8473 0.9205
No log 16.6923 434 0.8723 0.3902 0.8723 0.9340
No log 16.7692 436 0.8923 0.3902 0.8923 0.9446
No log 16.8462 438 0.8963 0.3552 0.8963 0.9467
No log 16.9231 440 0.8982 0.3596 0.8982 0.9477
No log 17.0 442 0.8932 0.3583 0.8932 0.9451
No log 17.0769 444 0.8912 0.3583 0.8912 0.9440
No log 17.1538 446 0.8853 0.3728 0.8853 0.9409
No log 17.2308 448 0.8847 0.4439 0.8847 0.9406
No log 17.3077 450 0.8778 0.4578 0.8778 0.9369
No log 17.3846 452 0.8463 0.3639 0.8463 0.9200
No log 17.4615 454 0.8650 0.3953 0.8650 0.9301
No log 17.5385 456 0.8801 0.4144 0.8801 0.9382
No log 17.6154 458 0.8872 0.3684 0.8872 0.9419
No log 17.6923 460 0.8871 0.3943 0.8871 0.9419
No log 17.7692 462 0.8919 0.3596 0.8919 0.9444
No log 17.8462 464 0.8946 0.3738 0.8946 0.9459
No log 17.9231 466 0.8916 0.3879 0.8916 0.9443
No log 18.0 468 0.9028 0.4297 0.9028 0.9502
No log 18.0769 470 0.9068 0.3855 0.9068 0.9522
No log 18.1538 472 0.8615 0.3738 0.8615 0.9282
No log 18.2308 474 0.8623 0.4142 0.8623 0.9286
No log 18.3077 476 0.9075 0.4775 0.9075 0.9526
No log 18.3846 478 0.9013 0.5014 0.9013 0.9494
No log 18.4615 480 0.8510 0.4476 0.8510 0.9225
No log 18.5385 482 0.8486 0.3838 0.8486 0.9212
No log 18.6154 484 0.8416 0.3796 0.8416 0.9174
No log 18.6923 486 0.8436 0.4476 0.8436 0.9185
No log 18.7692 488 0.8273 0.4476 0.8273 0.9096
No log 18.8462 490 0.8089 0.4181 0.8089 0.8994
No log 18.9231 492 0.8051 0.3987 0.8051 0.8973
No log 19.0 494 0.8174 0.4948 0.8174 0.9041
No log 19.0769 496 0.8331 0.5045 0.8331 0.9127
No log 19.1538 498 0.8520 0.4916 0.8520 0.9230
0.3866 19.2308 500 0.8380 0.4373 0.8380 0.9154
0.3866 19.3077 502 0.8351 0.4505 0.8351 0.9138
0.3866 19.3846 504 0.8443 0.4373 0.8443 0.9189
0.3866 19.4615 506 0.8555 0.4373 0.8555 0.9249
0.3866 19.5385 508 0.8485 0.4142 0.8485 0.9212
0.3866 19.6154 510 0.8394 0.4318 0.8394 0.9162
0.3866 19.6923 512 0.8395 0.3983 0.8395 0.9162

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k8_task2_organization

Finetuned
(4023)
this model