ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k7_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6847
  • Qwk: 0.4234
  • Mse: 0.6847
  • Rmse: 0.8275

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.08 2 2.5227 -0.0407 2.5227 1.5883
No log 0.16 4 1.1792 0.1565 1.1792 1.0859
No log 0.24 6 0.9680 -0.0622 0.9680 0.9839
No log 0.32 8 0.8298 -0.0027 0.8298 0.9110
No log 0.4 10 0.8954 0.0798 0.8954 0.9463
No log 0.48 12 0.9426 0.0428 0.9426 0.9709
No log 0.56 14 1.0014 0.0 1.0014 1.0007
No log 0.64 16 0.9426 0.0 0.9426 0.9709
No log 0.72 18 0.8572 0.0 0.8572 0.9259
No log 0.8 20 0.7974 0.0 0.7974 0.8930
No log 0.88 22 0.7344 0.0840 0.7344 0.8570
No log 0.96 24 0.7425 0.3084 0.7425 0.8617
No log 1.04 26 0.7208 0.1983 0.7208 0.8490
No log 1.12 28 0.6894 0.1617 0.6894 0.8303
No log 1.2 30 0.6930 0.0884 0.6930 0.8325
No log 1.28 32 0.7222 0.0481 0.7222 0.8498
No log 1.3600 34 0.7930 0.0937 0.7930 0.8905
No log 1.44 36 0.9042 0.1352 0.9042 0.9509
No log 1.52 38 0.8928 0.0509 0.8928 0.9449
No log 1.6 40 0.7673 0.0481 0.7673 0.8760
No log 1.6800 42 0.7131 0.1617 0.7131 0.8445
No log 1.76 44 0.7964 0.0327 0.7964 0.8924
No log 1.8400 46 0.7996 0.0840 0.7996 0.8942
No log 1.92 48 0.7862 0.0840 0.7862 0.8867
No log 2.0 50 0.7218 0.1287 0.7218 0.8496
No log 2.08 52 0.6928 -0.0027 0.6928 0.8324
No log 2.16 54 0.7167 0.0 0.7167 0.8466
No log 2.24 56 0.7121 0.0 0.7121 0.8439
No log 2.32 58 0.7353 0.1983 0.7353 0.8575
No log 2.4 60 0.7674 0.1236 0.7674 0.8760
No log 2.48 62 0.7694 0.0 0.7694 0.8772
No log 2.56 64 0.7843 0.0027 0.7843 0.8856
No log 2.64 66 0.8117 0.0522 0.8117 0.9010
No log 2.7200 68 0.9268 0.1973 0.9268 0.9627
No log 2.8 70 0.9023 0.1661 0.9023 0.9499
No log 2.88 72 0.7843 0.1267 0.7843 0.8856
No log 2.96 74 0.7577 0.1313 0.7577 0.8704
No log 3.04 76 0.8465 0.2244 0.8465 0.9200
No log 3.12 78 0.8561 0.2352 0.8561 0.9252
No log 3.2 80 0.7819 0.2950 0.7819 0.8843
No log 3.2800 82 0.7702 0.1456 0.7702 0.8776
No log 3.36 84 0.7762 0.1580 0.7762 0.8810
No log 3.44 86 0.7754 0.2368 0.7754 0.8806
No log 3.52 88 0.7533 0.3302 0.7533 0.8679
No log 3.6 90 0.8243 0.3243 0.8243 0.9079
No log 3.68 92 0.8114 0.3229 0.8114 0.9008
No log 3.76 94 0.8300 0.2484 0.8300 0.9111
No log 3.84 96 0.8448 0.2940 0.8448 0.9191
No log 3.92 98 0.8252 0.1960 0.8252 0.9084
No log 4.0 100 0.7820 0.2002 0.7820 0.8843
No log 4.08 102 0.7684 0.1353 0.7684 0.8766
No log 4.16 104 0.7856 0.2817 0.7856 0.8863
No log 4.24 106 0.7753 0.3261 0.7753 0.8805
No log 4.32 108 0.7504 0.3372 0.7504 0.8663
No log 4.4 110 0.7168 0.3166 0.7168 0.8466
No log 4.48 112 0.6895 0.3465 0.6895 0.8304
No log 4.5600 114 0.6751 0.3243 0.6751 0.8217
No log 4.64 116 0.6948 0.4243 0.6948 0.8336
No log 4.72 118 0.6791 0.4444 0.6791 0.8241
No log 4.8 120 0.6787 0.4660 0.6787 0.8239
No log 4.88 122 0.7018 0.4419 0.7018 0.8378
No log 4.96 124 0.7604 0.4147 0.7604 0.8720
No log 5.04 126 0.8114 0.2488 0.8114 0.9008
No log 5.12 128 0.8095 0.2053 0.8095 0.8997
No log 5.2 130 0.8161 0.3590 0.8161 0.9034
No log 5.28 132 0.8703 0.3159 0.8703 0.9329
No log 5.36 134 0.8081 0.2975 0.8081 0.8989
No log 5.44 136 0.7967 0.3034 0.7967 0.8926
No log 5.52 138 0.9691 0.2091 0.9691 0.9844
No log 5.6 140 0.9140 0.1497 0.9140 0.9560
No log 5.68 142 0.7822 0.2965 0.7822 0.8844
No log 5.76 144 0.8410 0.3034 0.8410 0.9171
No log 5.84 146 0.9301 0.2949 0.9301 0.9644
No log 5.92 148 0.8805 0.3157 0.8805 0.9383
No log 6.0 150 0.8534 0.1424 0.8534 0.9238
No log 6.08 152 0.8771 0.1528 0.8771 0.9365
No log 6.16 154 0.7933 0.1325 0.7933 0.8907
No log 6.24 156 0.7235 0.3864 0.7235 0.8506
No log 6.32 158 0.7209 0.4103 0.7209 0.8490
No log 6.4 160 0.7299 0.3980 0.7299 0.8544
No log 6.48 162 0.7994 0.4114 0.7994 0.8941
No log 6.5600 164 0.7673 0.3910 0.7673 0.8760
No log 6.64 166 0.7620 0.4328 0.7620 0.8729
No log 6.72 168 0.8154 0.2906 0.8154 0.9030
No log 6.8 170 0.7767 0.2016 0.7767 0.8813
No log 6.88 172 0.7932 0.4234 0.7932 0.8906
No log 6.96 174 0.9537 0.2394 0.9537 0.9766
No log 7.04 176 0.9661 0.2260 0.9661 0.9829
No log 7.12 178 0.8096 0.3261 0.8096 0.8998
No log 7.2 180 0.7431 0.3293 0.7431 0.8620
No log 7.28 182 0.8212 0.2193 0.8212 0.9062
No log 7.36 184 0.8764 0.1605 0.8764 0.9362
No log 7.44 186 0.8504 0.1924 0.8504 0.9222
No log 7.52 188 0.7990 0.1168 0.7990 0.8939
No log 7.6 190 0.7951 0.2078 0.7951 0.8917
No log 7.68 192 0.8362 0.3500 0.8362 0.9144
No log 7.76 194 0.8728 0.3409 0.8728 0.9343
No log 7.84 196 0.9144 0.2564 0.9144 0.9563
No log 7.92 198 0.8210 0.4134 0.8210 0.9061
No log 8.0 200 0.7830 0.4125 0.7830 0.8849
No log 8.08 202 0.8251 0.2741 0.8251 0.9083
No log 8.16 204 0.8268 0.2429 0.8268 0.9093
No log 8.24 206 0.8039 0.2862 0.8039 0.8966
No log 8.32 208 0.8287 0.3260 0.8287 0.9103
No log 8.4 210 0.8825 0.3146 0.8825 0.9394
No log 8.48 212 0.8859 0.3525 0.8859 0.9412
No log 8.56 214 0.9464 0.3060 0.9464 0.9728
No log 8.64 216 0.9302 0.2723 0.9302 0.9645
No log 8.72 218 0.8989 0.2968 0.8989 0.9481
No log 8.8 220 0.8286 0.3161 0.8286 0.9102
No log 8.88 222 0.8109 0.3861 0.8109 0.9005
No log 8.96 224 0.8085 0.3146 0.8085 0.8992
No log 9.04 226 0.8020 0.3746 0.8020 0.8956
No log 9.12 228 0.7821 0.3450 0.7821 0.8844
No log 9.2 230 0.7536 0.3918 0.7536 0.8681
No log 9.28 232 0.6952 0.3840 0.6952 0.8338
No log 9.36 234 0.7285 0.4451 0.7285 0.8535
No log 9.44 236 0.7908 0.4625 0.7908 0.8893
No log 9.52 238 0.7338 0.4408 0.7338 0.8566
No log 9.6 240 0.7076 0.3598 0.7076 0.8412
No log 9.68 242 0.7255 0.3768 0.7255 0.8518
No log 9.76 244 0.7221 0.3375 0.7221 0.8498
No log 9.84 246 0.7414 0.3701 0.7414 0.8610
No log 9.92 248 0.7929 0.4263 0.7929 0.8905
No log 10.0 250 0.8388 0.3543 0.8388 0.9159
No log 10.08 252 0.8125 0.2643 0.8125 0.9014
No log 10.16 254 0.7736 0.3441 0.7736 0.8795
No log 10.24 256 0.7880 0.4044 0.7880 0.8877
No log 10.32 258 0.7463 0.2689 0.7463 0.8639
No log 10.4 260 0.7377 0.3442 0.7377 0.8589
No log 10.48 262 0.7391 0.2999 0.7391 0.8597
No log 10.56 264 0.7746 0.3690 0.7746 0.8801
No log 10.64 266 0.8251 0.3173 0.8251 0.9084
No log 10.72 268 0.7925 0.3196 0.7925 0.8902
No log 10.8 270 0.8060 0.3083 0.8060 0.8977
No log 10.88 272 0.9166 0.3274 0.9166 0.9574
No log 10.96 274 0.8662 0.3333 0.8662 0.9307
No log 11.04 276 0.7373 0.3640 0.7373 0.8587
No log 11.12 278 0.7224 0.3762 0.7224 0.8499
No log 11.2 280 0.7608 0.3640 0.7608 0.8723
No log 11.28 282 0.8704 0.3194 0.8704 0.9329
No log 11.36 284 0.8886 0.3333 0.8886 0.9427
No log 11.44 286 0.7421 0.4212 0.7421 0.8614
No log 11.52 288 0.6815 0.4278 0.6815 0.8255
No log 11.6 290 0.6852 0.4234 0.6852 0.8278
No log 11.68 292 0.7277 0.3930 0.7277 0.8531
No log 11.76 294 0.6716 0.4562 0.6716 0.8195
No log 11.84 296 0.6478 0.4581 0.6478 0.8049
No log 11.92 298 0.6425 0.4342 0.6425 0.8016
No log 12.0 300 0.6533 0.5133 0.6533 0.8083
No log 12.08 302 0.7351 0.4916 0.7351 0.8574
No log 12.16 304 0.7540 0.4542 0.7540 0.8684
No log 12.24 306 0.7281 0.4523 0.7281 0.8533
No log 12.32 308 0.7259 0.4073 0.7259 0.8520
No log 12.4 310 0.7255 0.2973 0.7255 0.8518
No log 12.48 312 0.7212 0.2634 0.7212 0.8492
No log 12.56 314 0.7379 0.5081 0.7379 0.8590
No log 12.64 316 0.7887 0.4315 0.7887 0.8881
No log 12.72 318 0.7345 0.4464 0.7345 0.8570
No log 12.8 320 0.6923 0.3472 0.6923 0.8320
No log 12.88 322 0.7280 0.3455 0.7280 0.8532
No log 12.96 324 0.7216 0.4196 0.7216 0.8495
No log 13.04 326 0.8100 0.4550 0.8100 0.9000
No log 13.12 328 0.8011 0.4212 0.8011 0.8950
No log 13.2 330 0.7152 0.4125 0.7152 0.8457
No log 13.28 332 0.7069 0.3311 0.7069 0.8408
No log 13.36 334 0.7081 0.3311 0.7081 0.8415
No log 13.44 336 0.7065 0.3034 0.7065 0.8405
No log 13.52 338 0.7107 0.3738 0.7107 0.8430
No log 13.6 340 0.7151 0.2453 0.7151 0.8457
No log 13.68 342 0.7319 0.2867 0.7319 0.8555
No log 13.76 344 0.7185 0.2867 0.7185 0.8477
No log 13.84 346 0.6886 0.4419 0.6886 0.8298
No log 13.92 348 0.7295 0.4352 0.7295 0.8541
No log 14.0 350 0.7199 0.4352 0.7199 0.8485
No log 14.08 352 0.6973 0.4267 0.6973 0.8351
No log 14.16 354 0.6785 0.3551 0.6785 0.8237
No log 14.24 356 0.6846 0.3551 0.6846 0.8274
No log 14.32 358 0.7130 0.3990 0.7130 0.8444
No log 14.4 360 0.7493 0.4352 0.7493 0.8656
No log 14.48 362 0.7888 0.3746 0.7888 0.8882
No log 14.56 364 0.8293 0.3653 0.8293 0.9106
No log 14.64 366 0.8140 0.3985 0.8140 0.9022
No log 14.72 368 0.7447 0.4234 0.7447 0.8630
No log 14.8 370 0.7165 0.3311 0.7165 0.8464
No log 14.88 372 0.7034 0.4535 0.7034 0.8387
No log 14.96 374 0.7193 0.4493 0.7193 0.8481
No log 15.04 376 0.7360 0.3841 0.7360 0.8579
No log 15.12 378 0.7131 0.4743 0.7131 0.8444
No log 15.2 380 0.7017 0.4681 0.7017 0.8377
No log 15.28 382 0.7042 0.4418 0.7042 0.8391
No log 15.36 384 0.7107 0.4438 0.7107 0.8431
No log 15.44 386 0.7036 0.4116 0.7036 0.8388
No log 15.52 388 0.7191 0.4434 0.7191 0.8480
No log 15.6 390 0.7236 0.4455 0.7236 0.8506
No log 15.68 392 0.7316 0.3235 0.7316 0.8554
No log 15.76 394 0.7334 0.2511 0.7334 0.8564
No log 15.84 396 0.7255 0.3961 0.7255 0.8517
No log 15.92 398 0.7511 0.4618 0.7511 0.8667
No log 16.0 400 0.8104 0.3409 0.8104 0.9002
No log 16.08 402 0.7619 0.4212 0.7619 0.8729
No log 16.16 404 0.6978 0.3504 0.6978 0.8353
No log 16.24 406 0.6849 0.3577 0.6849 0.8276
No log 16.32 408 0.6752 0.4027 0.6752 0.8217
No log 16.4 410 0.6750 0.3945 0.6750 0.8216
No log 16.48 412 0.6786 0.3945 0.6786 0.8238
No log 16.56 414 0.6804 0.3945 0.6804 0.8249
No log 16.64 416 0.6809 0.3945 0.6809 0.8252
No log 16.72 418 0.7027 0.3788 0.7027 0.8383
No log 16.8 420 0.7231 0.4451 0.7231 0.8503
No log 16.88 422 0.7585 0.4093 0.7585 0.8709
No log 16.96 424 0.8096 0.4180 0.8096 0.8998
No log 17.04 426 0.7331 0.4072 0.7331 0.8562
No log 17.12 428 0.6530 0.5271 0.6530 0.8081
No log 17.2 430 0.6493 0.4618 0.6493 0.8058
No log 17.28 432 0.6683 0.4855 0.6683 0.8175
No log 17.36 434 0.6801 0.5149 0.6801 0.8247
No log 17.44 436 0.6256 0.4816 0.6256 0.7909
No log 17.52 438 0.6071 0.4768 0.6071 0.7792
No log 17.6 440 0.6042 0.4747 0.6042 0.7773
No log 17.68 442 0.6560 0.4684 0.6560 0.8099
No log 17.76 444 0.8019 0.4133 0.8019 0.8955
No log 17.84 446 0.8680 0.3928 0.8680 0.9317
No log 17.92 448 0.7798 0.4450 0.7798 0.8831
No log 18.0 450 0.6485 0.5017 0.6485 0.8053
No log 18.08 452 0.6198 0.4027 0.6198 0.7873
No log 18.16 454 0.6316 0.4526 0.6316 0.7948
No log 18.24 456 0.6324 0.4229 0.6324 0.7952
No log 18.32 458 0.6908 0.4582 0.6908 0.8311
No log 18.4 460 0.7834 0.3869 0.7834 0.8851
No log 18.48 462 0.7368 0.4582 0.7368 0.8584
No log 18.56 464 0.6664 0.4234 0.6664 0.8163
No log 18.64 466 0.6579 0.3603 0.6579 0.8111
No log 18.72 468 0.6487 0.3984 0.6487 0.8054
No log 18.8 470 0.6525 0.4618 0.6525 0.8078
No log 18.88 472 0.6643 0.4855 0.6643 0.8151
No log 18.96 474 0.6391 0.4855 0.6391 0.7994
No log 19.04 476 0.6262 0.4253 0.6262 0.7913
No log 19.12 478 0.6432 0.3886 0.6432 0.8020
No log 19.2 480 0.6436 0.3467 0.6436 0.8023
No log 19.28 482 0.6401 0.4660 0.6401 0.8001
No log 19.36 484 0.7102 0.3699 0.7102 0.8427
No log 19.44 486 0.7455 0.3630 0.7455 0.8634
No log 19.52 488 0.7342 0.3770 0.7342 0.8568
No log 19.6 490 0.6823 0.5036 0.6823 0.8260
No log 19.68 492 0.6726 0.2867 0.6726 0.8201
No log 19.76 494 0.6616 0.3198 0.6616 0.8134
No log 19.84 496 0.6561 0.4875 0.6561 0.8100
No log 19.92 498 0.6573 0.4597 0.6573 0.8107
0.3111 20.0 500 0.6604 0.4352 0.6604 0.8127
0.3111 20.08 502 0.6477 0.5036 0.6477 0.8048
0.3111 20.16 504 0.6539 0.3754 0.6539 0.8086
0.3111 20.24 506 0.6812 0.3051 0.6812 0.8254
0.3111 20.32 508 0.6630 0.3703 0.6630 0.8142
0.3111 20.4 510 0.6393 0.4419 0.6393 0.7996
0.3111 20.48 512 0.6419 0.5131 0.6419 0.8012
0.3111 20.56 514 0.6403 0.3352 0.6403 0.8002
0.3111 20.64 516 0.6436 0.4378 0.6436 0.8022
0.3111 20.72 518 0.6494 0.3754 0.6494 0.8059
0.3111 20.8 520 0.6589 0.4229 0.6589 0.8117
0.3111 20.88 522 0.6631 0.3100 0.6631 0.8143
0.3111 20.96 524 0.6616 0.3366 0.6616 0.8134
0.3111 21.04 526 0.6847 0.4234 0.6847 0.8275

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k7_task7_organization

Finetuned
(4019)
this model