ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k11_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8170
  • Qwk: 0.3301
  • Mse: 0.8170
  • Rmse: 0.9039

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0385 2 4.0034 -0.0232 4.0034 2.0009
No log 0.0769 4 2.4408 0.0014 2.4408 1.5623
No log 0.1154 6 2.0145 0.0119 2.0145 1.4193
No log 0.1538 8 1.3068 0.2199 1.3068 1.1432
No log 0.1923 10 1.0860 0.2094 1.0860 1.0421
No log 0.2308 12 1.0529 0.1989 1.0529 1.0261
No log 0.2692 14 1.0617 0.2140 1.0617 1.0304
No log 0.3077 16 1.0699 0.2341 1.0699 1.0344
No log 0.3462 18 1.2528 0.1564 1.2528 1.1193
No log 0.3846 20 1.2655 0.1057 1.2655 1.1250
No log 0.4231 22 1.1667 0.2196 1.1667 1.0801
No log 0.4615 24 1.0282 0.1713 1.0282 1.0140
No log 0.5 26 1.0187 0.1532 1.0187 1.0093
No log 0.5385 28 1.0701 0.2416 1.0701 1.0344
No log 0.5769 30 1.1895 0.1711 1.1895 1.0906
No log 0.6154 32 1.1995 0.1711 1.1995 1.0952
No log 0.6538 34 1.0443 0.1603 1.0443 1.0219
No log 0.6923 36 0.9649 0.3915 0.9649 0.9823
No log 0.7308 38 0.9157 0.3721 0.9157 0.9569
No log 0.7692 40 0.8974 0.3569 0.8974 0.9473
No log 0.8077 42 1.0305 0.3977 1.0305 1.0151
No log 0.8462 44 1.5196 0.2412 1.5196 1.2327
No log 0.8846 46 1.4412 0.2399 1.4412 1.2005
No log 0.9231 48 0.9751 0.3646 0.9751 0.9875
No log 0.9615 50 0.9234 0.3637 0.9234 0.9610
No log 1.0 52 0.9269 0.3494 0.9269 0.9627
No log 1.0385 54 0.9055 0.2812 0.9055 0.9516
No log 1.0769 56 0.9838 0.2782 0.9838 0.9919
No log 1.1154 58 1.0486 0.2988 1.0486 1.0240
No log 1.1538 60 1.4679 0.1880 1.4679 1.2116
No log 1.1923 62 1.5374 0.2188 1.5374 1.2399
No log 1.2308 64 1.1832 0.2215 1.1832 1.0878
No log 1.2692 66 0.9056 0.3151 0.9056 0.9516
No log 1.3077 68 0.9630 0.3637 0.9630 0.9813
No log 1.3462 70 0.9382 0.3272 0.9382 0.9686
No log 1.3846 72 0.9300 0.3458 0.9300 0.9644
No log 1.4231 74 1.1969 0.4100 1.1969 1.0940
No log 1.4615 76 1.2119 0.4100 1.2119 1.1009
No log 1.5 78 0.9774 0.3187 0.9774 0.9886
No log 1.5385 80 0.8607 0.3089 0.8607 0.9277
No log 1.5769 82 0.9460 0.3617 0.9460 0.9726
No log 1.6154 84 0.9677 0.3103 0.9677 0.9837
No log 1.6538 86 0.9507 0.3207 0.9507 0.9750
No log 1.6923 88 0.9200 0.3044 0.9200 0.9591
No log 1.7308 90 0.8696 0.3536 0.8696 0.9325
No log 1.7692 92 0.8743 0.3544 0.8743 0.9351
No log 1.8077 94 0.8942 0.3257 0.8942 0.9456
No log 1.8462 96 0.9060 0.3496 0.9060 0.9518
No log 1.8846 98 0.9812 0.2842 0.9812 0.9906
No log 1.9231 100 1.3123 0.1616 1.3123 1.1456
No log 1.9615 102 1.6396 0.0529 1.6396 1.2805
No log 2.0 104 1.4156 0.1192 1.4156 1.1898
No log 2.0385 106 1.0074 0.3124 1.0074 1.0037
No log 2.0769 108 0.9328 0.4822 0.9328 0.9658
No log 2.1154 110 1.0684 0.3283 1.0684 1.0336
No log 2.1538 112 1.0180 0.3659 1.0180 1.0089
No log 2.1923 114 0.9403 0.3671 0.9403 0.9697
No log 2.2308 116 1.0544 0.2705 1.0544 1.0268
No log 2.2692 118 1.0685 0.2747 1.0685 1.0337
No log 2.3077 120 1.0512 0.3950 1.0512 1.0253
No log 2.3462 122 1.0647 0.3437 1.0647 1.0318
No log 2.3846 124 1.0569 0.2562 1.0569 1.0281
No log 2.4231 126 1.2021 0.2935 1.2021 1.0964
No log 2.4615 128 1.1833 0.3268 1.1833 1.0878
No log 2.5 130 1.0240 0.3042 1.0240 1.0119
No log 2.5385 132 0.9014 0.3129 0.9014 0.9494
No log 2.5769 134 0.8693 0.3603 0.8693 0.9324
No log 2.6154 136 0.8686 0.4313 0.8686 0.9320
No log 2.6538 138 0.8780 0.4661 0.8780 0.9370
No log 2.6923 140 0.9805 0.3573 0.9805 0.9902
No log 2.7308 142 1.0312 0.3231 1.0312 1.0155
No log 2.7692 144 0.9392 0.3773 0.9392 0.9691
No log 2.8077 146 0.8641 0.5934 0.8641 0.9296
No log 2.8462 148 0.9977 0.4850 0.9977 0.9988
No log 2.8846 150 1.0046 0.4642 1.0046 1.0023
No log 2.9231 152 0.8675 0.5934 0.8675 0.9314
No log 2.9615 154 0.9210 0.3977 0.9210 0.9597
No log 3.0 156 1.1516 0.3215 1.1516 1.0731
No log 3.0385 158 1.2183 0.3568 1.2183 1.1037
No log 3.0769 160 1.0982 0.3391 1.0982 1.0479
No log 3.1154 162 0.9352 0.3648 0.9352 0.9671
No log 3.1538 164 0.8509 0.3596 0.8509 0.9225
No log 3.1923 166 0.8878 0.4608 0.8878 0.9422
No log 3.2308 168 0.8949 0.4990 0.8949 0.9460
No log 3.2692 170 0.8519 0.4450 0.8519 0.9230
No log 3.3077 172 0.8691 0.4123 0.8691 0.9323
No log 3.3462 174 0.9366 0.4241 0.9366 0.9678
No log 3.3846 176 0.9380 0.4275 0.9380 0.9685
No log 3.4231 178 0.9125 0.4751 0.9125 0.9552
No log 3.4615 180 0.8873 0.4662 0.8873 0.9419
No log 3.5 182 0.8856 0.4662 0.8856 0.9410
No log 3.5385 184 0.8740 0.4550 0.8740 0.9349
No log 3.5769 186 0.8801 0.4288 0.8801 0.9381
No log 3.6154 188 0.9173 0.3551 0.9173 0.9578
No log 3.6538 190 0.9176 0.3434 0.9176 0.9579
No log 3.6923 192 0.8649 0.4417 0.8649 0.9300
No log 3.7308 194 0.9012 0.4734 0.9012 0.9493
No log 3.7692 196 0.9478 0.4469 0.9478 0.9735
No log 3.8077 198 0.8991 0.4541 0.8991 0.9482
No log 3.8462 200 0.8684 0.4645 0.8684 0.9319
No log 3.8846 202 1.0168 0.4014 1.0168 1.0084
No log 3.9231 204 1.1310 0.3744 1.1310 1.0635
No log 3.9615 206 1.0464 0.3496 1.0464 1.0229
No log 4.0 208 0.9382 0.4549 0.9382 0.9686
No log 4.0385 210 0.8596 0.3958 0.8596 0.9271
No log 4.0769 212 0.8562 0.3800 0.8562 0.9253
No log 4.1154 214 0.8312 0.4313 0.8312 0.9117
No log 4.1538 216 0.8328 0.4247 0.8328 0.9126
No log 4.1923 218 0.8797 0.4234 0.8797 0.9379
No log 4.2308 220 0.8402 0.4106 0.8402 0.9166
No log 4.2692 222 0.8065 0.4297 0.8065 0.8981
No log 4.3077 224 0.8162 0.5046 0.8162 0.9034
No log 4.3462 226 0.8255 0.3781 0.8255 0.9085
No log 4.3846 228 0.8550 0.3465 0.8550 0.9246
No log 4.4231 230 0.8592 0.3744 0.8592 0.9269
No log 4.4615 232 0.8564 0.3931 0.8564 0.9254
No log 4.5 234 0.8702 0.4413 0.8702 0.9329
No log 4.5385 236 0.8664 0.4691 0.8664 0.9308
No log 4.5769 238 0.9157 0.5313 0.9157 0.9569
No log 4.6154 240 0.9043 0.5313 0.9043 0.9510
No log 4.6538 242 0.8584 0.5287 0.8584 0.9265
No log 4.6923 244 0.8641 0.3959 0.8641 0.9296
No log 4.7308 246 0.8977 0.3596 0.8977 0.9475
No log 4.7692 248 0.8935 0.3822 0.8935 0.9452
No log 4.8077 250 0.8696 0.3258 0.8696 0.9325
No log 4.8462 252 0.9173 0.3265 0.9173 0.9578
No log 4.8846 254 0.9155 0.3265 0.9155 0.9568
No log 4.9231 256 0.8901 0.3222 0.8901 0.9434
No log 4.9615 258 0.8763 0.3878 0.8763 0.9361
No log 5.0 260 0.9123 0.3285 0.9123 0.9551
No log 5.0385 262 0.9420 0.3822 0.9420 0.9705
No log 5.0769 264 0.8981 0.3821 0.8981 0.9477
No log 5.1154 266 0.8798 0.3878 0.8798 0.9380
No log 5.1538 268 0.9090 0.4544 0.9090 0.9534
No log 5.1923 270 0.9159 0.3744 0.9159 0.9570
No log 5.2308 272 1.0030 0.3380 1.0030 1.0015
No log 5.2692 274 1.1454 0.2907 1.1454 1.0702
No log 5.3077 276 1.3438 0.3725 1.3438 1.1592
No log 5.3462 278 1.3613 0.3484 1.3613 1.1667
No log 5.3846 280 1.2050 0.3578 1.2050 1.0977
No log 5.4231 282 1.0307 0.3724 1.0307 1.0152
No log 5.4615 284 0.9045 0.2671 0.9045 0.9511
No log 5.5 286 0.9062 0.3133 0.9062 0.9519
No log 5.5385 288 0.9262 0.4258 0.9262 0.9624
No log 5.5769 290 0.9070 0.3820 0.9070 0.9524
No log 5.6154 292 0.9182 0.3657 0.9182 0.9582
No log 5.6538 294 0.9439 0.3029 0.9439 0.9715
No log 5.6923 296 1.0427 0.3099 1.0427 1.0211
No log 5.7308 298 1.0862 0.3681 1.0862 1.0422
No log 5.7692 300 1.0018 0.2726 1.0018 1.0009
No log 5.8077 302 0.8724 0.3285 0.8724 0.9340
No log 5.8462 304 0.8409 0.3821 0.8409 0.9170
No log 5.8846 306 0.8309 0.3821 0.8309 0.9115
No log 5.9231 308 0.8298 0.3590 0.8298 0.9109
No log 5.9615 310 0.8114 0.3976 0.8114 0.9008
No log 6.0 312 0.8242 0.4338 0.8242 0.9078
No log 6.0385 314 0.8371 0.4355 0.8371 0.9149
No log 6.0769 316 0.8136 0.3837 0.8136 0.9020
No log 6.1154 318 0.8209 0.3089 0.8209 0.9060
No log 6.1538 320 0.9288 0.2674 0.9288 0.9637
No log 6.1923 322 0.9952 0.3483 0.9952 0.9976
No log 6.2308 324 0.9653 0.3363 0.9653 0.9825
No log 6.2692 326 0.9469 0.3921 0.9469 0.9731
No log 6.3077 328 0.8564 0.3822 0.8564 0.9254
No log 6.3462 330 0.8028 0.4490 0.8028 0.8960
No log 6.3846 332 0.7907 0.4995 0.7907 0.8892
No log 6.4231 334 0.8070 0.4483 0.8070 0.8983
No log 6.4615 336 0.8396 0.3644 0.8396 0.9163
No log 6.5 338 0.8505 0.4192 0.8505 0.9222
No log 6.5385 340 0.8448 0.4057 0.8448 0.9191
No log 6.5769 342 0.8256 0.3822 0.8256 0.9086
No log 6.6154 344 0.8339 0.2671 0.8339 0.9132
No log 6.6538 346 0.8724 0.2771 0.8724 0.9340
No log 6.6923 348 0.9214 0.3643 0.9214 0.9599
No log 6.7308 350 0.9102 0.3643 0.9102 0.9540
No log 6.7692 352 0.8836 0.3643 0.8836 0.9400
No log 6.8077 354 0.8475 0.3285 0.8475 0.9206
No log 6.8462 356 0.8550 0.3687 0.8550 0.9247
No log 6.8846 358 0.8686 0.3822 0.8686 0.9320
No log 6.9231 360 0.8830 0.3687 0.8830 0.9397
No log 6.9615 362 0.8705 0.3285 0.8705 0.9330
No log 7.0 364 0.8540 0.3335 0.8540 0.9241
No log 7.0385 366 0.8929 0.4210 0.8929 0.9449
No log 7.0769 368 0.8786 0.3814 0.8786 0.9373
No log 7.1154 370 0.8400 0.3323 0.8400 0.9165
No log 7.1538 372 0.8258 0.3112 0.8258 0.9087
No log 7.1923 374 0.8437 0.3393 0.8437 0.9185
No log 7.2308 376 0.8531 0.4180 0.8531 0.9237
No log 7.2692 378 0.8402 0.4310 0.8402 0.9166
No log 7.3077 380 0.7920 0.4344 0.7920 0.8899
No log 7.3462 382 0.7742 0.3941 0.7742 0.8799
No log 7.3846 384 0.7700 0.3652 0.7700 0.8775
No log 7.4231 386 0.7829 0.3652 0.7829 0.8848
No log 7.4615 388 0.7846 0.3652 0.7846 0.8858
No log 7.5 390 0.7864 0.4086 0.7864 0.8868
No log 7.5385 392 0.7942 0.4230 0.7942 0.8912
No log 7.5769 394 0.7946 0.4105 0.7946 0.8914
No log 7.6154 396 0.7970 0.3896 0.7970 0.8927
No log 7.6538 398 0.7909 0.4138 0.7909 0.8893
No log 7.6923 400 0.7924 0.4524 0.7924 0.8901
No log 7.7308 402 0.8267 0.3958 0.8267 0.9092
No log 7.7692 404 0.8293 0.3682 0.8293 0.9106
No log 7.8077 406 0.7941 0.3272 0.7941 0.8911
No log 7.8462 408 0.7734 0.4381 0.7734 0.8794
No log 7.8846 410 0.7831 0.4557 0.7831 0.8849
No log 7.9231 412 0.8327 0.4739 0.8327 0.9125
No log 7.9615 414 0.9184 0.3812 0.9184 0.9583
No log 8.0 416 0.9159 0.3812 0.9159 0.9570
No log 8.0385 418 0.9084 0.3431 0.9084 0.9531
No log 8.0769 420 0.8424 0.4238 0.8424 0.9178
No log 8.1154 422 0.8143 0.4944 0.8143 0.9024
No log 8.1538 424 0.8042 0.4012 0.8042 0.8968
No log 8.1923 426 0.8096 0.3932 0.8096 0.8998
No log 8.2308 428 0.8160 0.4163 0.8160 0.9033
No log 8.2692 430 0.8183 0.4133 0.8183 0.9046
No log 8.3077 432 0.8165 0.4133 0.8165 0.9036
No log 8.3462 434 0.8126 0.4133 0.8126 0.9014
No log 8.3846 436 0.8237 0.4422 0.8237 0.9076
No log 8.4231 438 0.8667 0.4484 0.8667 0.9310
No log 8.4615 440 0.9044 0.4389 0.9044 0.9510
No log 8.5 442 0.8742 0.4755 0.8742 0.9350
No log 8.5385 444 0.8263 0.4838 0.8263 0.9090
No log 8.5769 446 0.8249 0.4279 0.8249 0.9083
No log 8.6154 448 0.8817 0.3634 0.8817 0.9390
No log 8.6538 450 0.9106 0.2791 0.9106 0.9542
No log 8.6923 452 0.9278 0.2577 0.9278 0.9632
No log 8.7308 454 0.9312 0.2577 0.9312 0.9650
No log 8.7692 456 0.9546 0.3804 0.9546 0.9771
No log 8.8077 458 0.9909 0.3691 0.9909 0.9954
No log 8.8462 460 0.9649 0.3710 0.9649 0.9823
No log 8.8846 462 0.9372 0.3958 0.9372 0.9681
No log 8.9231 464 0.8793 0.3445 0.8793 0.9377
No log 8.9615 466 0.8485 0.4422 0.8485 0.9211
No log 9.0 468 0.8516 0.3816 0.8516 0.9228
No log 9.0385 470 0.8503 0.4186 0.8503 0.9221
No log 9.0769 472 0.8659 0.3293 0.8659 0.9305
No log 9.1154 474 0.8920 0.3583 0.8920 0.9445
No log 9.1538 476 0.9492 0.3383 0.9492 0.9743
No log 9.1923 478 0.9244 0.2887 0.9244 0.9614
No log 9.2308 480 0.8692 0.3465 0.8692 0.9323
No log 9.2692 482 0.8501 0.3496 0.8501 0.9220
No log 9.3077 484 0.8504 0.3636 0.8504 0.9222
No log 9.3462 486 0.8566 0.3590 0.8566 0.9255
No log 9.3846 488 0.8656 0.3569 0.8656 0.9304
No log 9.4231 490 0.8748 0.3548 0.8748 0.9353
No log 9.4615 492 0.8764 0.2988 0.8764 0.9361
No log 9.5 494 0.8526 0.3590 0.8526 0.9234
No log 9.5385 496 0.8436 0.4019 0.8436 0.9185
No log 9.5769 498 0.8267 0.4156 0.8267 0.9093
0.2897 9.6154 500 0.8132 0.4810 0.8132 0.9018
0.2897 9.6538 502 0.8061 0.4675 0.8061 0.8978
0.2897 9.6923 504 0.8002 0.4124 0.8002 0.8945
0.2897 9.7308 506 0.7990 0.4143 0.7990 0.8939
0.2897 9.7692 508 0.8078 0.3713 0.8078 0.8988
0.2897 9.8077 510 0.8170 0.3301 0.8170 0.9039

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k11_task5_organization

Finetuned
(4019)
this model