ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k15_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8154
  • Qwk: 0.6517
  • Mse: 0.8154
  • Rmse: 0.9030

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0267 2 5.4219 -0.0290 5.4219 2.3285
No log 0.0533 4 3.0978 0.0548 3.0978 1.7601
No log 0.08 6 2.2219 0.0077 2.2219 1.4906
No log 0.1067 8 1.4819 0.1051 1.4819 1.2173
No log 0.1333 10 1.2387 0.3249 1.2387 1.1130
No log 0.16 12 1.0836 0.2351 1.0836 1.0409
No log 0.1867 14 1.1270 0.2255 1.1270 1.0616
No log 0.2133 16 1.1284 0.2255 1.1284 1.0622
No log 0.24 18 1.0956 0.1298 1.0956 1.0467
No log 0.2667 20 1.1011 0.1510 1.1011 1.0493
No log 0.2933 22 1.1087 0.2115 1.1087 1.0530
No log 0.32 24 1.2008 0.1581 1.2008 1.0958
No log 0.3467 26 1.2578 0.0990 1.2578 1.1215
No log 0.3733 28 1.0780 0.2013 1.0780 1.0383
No log 0.4 30 1.0453 0.2136 1.0453 1.0224
No log 0.4267 32 1.0796 0.2926 1.0796 1.0390
No log 0.4533 34 1.1769 0.3064 1.1769 1.0849
No log 0.48 36 1.1974 0.3147 1.1974 1.0943
No log 0.5067 38 1.0706 0.3976 1.0706 1.0347
No log 0.5333 40 1.0705 0.4354 1.0705 1.0346
No log 0.56 42 1.0832 0.3821 1.0832 1.0408
No log 0.5867 44 1.0815 0.3669 1.0815 1.0399
No log 0.6133 46 1.0870 0.3316 1.0870 1.0426
No log 0.64 48 1.0823 0.2866 1.0823 1.0403
No log 0.6667 50 1.0826 0.2386 1.0826 1.0405
No log 0.6933 52 1.0854 0.2212 1.0854 1.0418
No log 0.72 54 1.0465 0.2129 1.0465 1.0230
No log 0.7467 56 1.1061 0.3800 1.1061 1.0517
No log 0.7733 58 1.2422 0.1716 1.2422 1.1145
No log 0.8 60 1.2621 0.1252 1.2621 1.1234
No log 0.8267 62 1.2416 0.1933 1.2416 1.1143
No log 0.8533 64 1.1712 0.2701 1.1712 1.0822
No log 0.88 66 1.1499 0.3361 1.1499 1.0723
No log 0.9067 68 1.2887 0.3126 1.2887 1.1352
No log 0.9333 70 1.2299 0.4163 1.2299 1.1090
No log 0.96 72 1.0519 0.4035 1.0519 1.0256
No log 0.9867 74 0.9968 0.4214 0.9968 0.9984
No log 1.0133 76 0.9993 0.4710 0.9993 0.9997
No log 1.04 78 0.9719 0.4924 0.9719 0.9859
No log 1.0667 80 0.9576 0.4935 0.9576 0.9786
No log 1.0933 82 0.9159 0.5426 0.9159 0.9570
No log 1.12 84 0.9336 0.5518 0.9336 0.9662
No log 1.1467 86 1.0643 0.5075 1.0643 1.0317
No log 1.1733 88 1.1084 0.4583 1.1084 1.0528
No log 1.2 90 1.0845 0.5038 1.0845 1.0414
No log 1.2267 92 0.9273 0.4542 0.9273 0.9630
No log 1.2533 94 0.8174 0.4676 0.8174 0.9041
No log 1.28 96 0.9536 0.4691 0.9536 0.9765
No log 1.3067 98 1.0118 0.4624 1.0118 1.0059
No log 1.3333 100 0.8547 0.5509 0.8547 0.9245
No log 1.3600 102 0.7899 0.5578 0.7899 0.8888
No log 1.3867 104 0.8490 0.6134 0.8490 0.9214
No log 1.4133 106 1.0529 0.4828 1.0529 1.0261
No log 1.44 108 1.0751 0.4840 1.0751 1.0369
No log 1.4667 110 0.9827 0.5392 0.9827 0.9913
No log 1.4933 112 0.9394 0.5470 0.9394 0.9692
No log 1.52 114 1.0210 0.4848 1.0210 1.0104
No log 1.5467 116 1.0370 0.4879 1.0370 1.0183
No log 1.5733 118 1.1324 0.4177 1.1324 1.0642
No log 1.6 120 1.3455 0.3241 1.3455 1.1600
No log 1.6267 122 1.4387 0.3612 1.4387 1.1995
No log 1.6533 124 1.4433 0.3887 1.4433 1.2014
No log 1.6800 126 1.3579 0.3966 1.3579 1.1653
No log 1.7067 128 1.2213 0.4437 1.2213 1.1051
No log 1.7333 130 1.0231 0.4962 1.0231 1.0115
No log 1.76 132 0.8561 0.4838 0.8561 0.9252
No log 1.7867 134 0.9759 0.4630 0.9759 0.9879
No log 1.8133 136 0.9956 0.4566 0.9956 0.9978
No log 1.8400 138 0.8562 0.5076 0.8562 0.9253
No log 1.8667 140 1.1133 0.5036 1.1133 1.0551
No log 1.8933 142 1.3100 0.4262 1.3100 1.1446
No log 1.92 144 1.3623 0.4163 1.3623 1.1672
No log 1.9467 146 1.4264 0.3901 1.4264 1.1943
No log 1.9733 148 1.3823 0.4113 1.3823 1.1757
No log 2.0 150 1.1573 0.4751 1.1573 1.0758
No log 2.0267 152 0.9886 0.5128 0.9886 0.9943
No log 2.0533 154 0.9817 0.5128 0.9817 0.9908
No log 2.08 156 1.0954 0.4841 1.0954 1.0466
No log 2.1067 158 1.2485 0.4575 1.2485 1.1174
No log 2.1333 160 1.2567 0.4402 1.2567 1.1210
No log 2.16 162 1.1686 0.4205 1.1686 1.0810
No log 2.1867 164 1.1766 0.4494 1.1766 1.0847
No log 2.2133 166 1.1197 0.4843 1.1197 1.0581
No log 2.24 168 1.0137 0.4886 1.0137 1.0068
No log 2.2667 170 0.8993 0.5811 0.8993 0.9483
No log 2.2933 172 0.8337 0.5646 0.8337 0.9131
No log 2.32 174 0.8105 0.5501 0.8105 0.9003
No log 2.3467 176 0.8188 0.5660 0.8188 0.9049
No log 2.3733 178 0.8291 0.6038 0.8291 0.9106
No log 2.4 180 0.9570 0.5849 0.9570 0.9783
No log 2.4267 182 1.2832 0.4419 1.2832 1.1328
No log 2.4533 184 1.3968 0.3967 1.3968 1.1818
No log 2.48 186 1.3017 0.3863 1.3017 1.1409
No log 2.5067 188 1.4026 0.3916 1.4026 1.1843
No log 2.5333 190 1.6154 0.3541 1.6154 1.2710
No log 2.56 192 1.6371 0.3598 1.6371 1.2795
No log 2.5867 194 1.3547 0.4310 1.3547 1.1639
No log 2.6133 196 1.0496 0.4805 1.0496 1.0245
No log 2.64 198 0.8834 0.6002 0.8834 0.9399
No log 2.6667 200 0.8588 0.6011 0.8588 0.9267
No log 2.6933 202 0.8948 0.5728 0.8948 0.9459
No log 2.7200 204 0.9001 0.5588 0.9001 0.9488
No log 2.7467 206 0.8165 0.6030 0.8165 0.9036
No log 2.7733 208 0.8376 0.5993 0.8376 0.9152
No log 2.8 210 1.0367 0.5225 1.0367 1.0182
No log 2.8267 212 1.1893 0.4642 1.1893 1.0905
No log 2.8533 214 1.2299 0.4381 1.2299 1.1090
No log 2.88 216 1.1531 0.5158 1.1531 1.0738
No log 2.9067 218 1.0907 0.5386 1.0907 1.0444
No log 2.9333 220 1.1481 0.5261 1.1481 1.0715
No log 2.96 222 1.2043 0.4982 1.2043 1.0974
No log 2.9867 224 1.0845 0.5550 1.0845 1.0414
No log 3.0133 226 0.9419 0.5992 0.9419 0.9705
No log 3.04 228 0.8802 0.6546 0.8802 0.9382
No log 3.0667 230 0.8614 0.6281 0.8614 0.9281
No log 3.0933 232 0.9378 0.5786 0.9378 0.9684
No log 3.12 234 1.1605 0.4838 1.1605 1.0772
No log 3.1467 236 1.3780 0.4245 1.3780 1.1739
No log 3.1733 238 1.3787 0.4245 1.3787 1.1742
No log 3.2 240 1.1199 0.5077 1.1199 1.0583
No log 3.2267 242 0.8667 0.6227 0.8667 0.9310
No log 3.2533 244 0.7665 0.6333 0.7665 0.8755
No log 3.2800 246 0.7587 0.6479 0.7587 0.8710
No log 3.3067 248 0.7987 0.6450 0.7987 0.8937
No log 3.3333 250 0.9230 0.5831 0.9230 0.9607
No log 3.36 252 1.1372 0.5480 1.1372 1.0664
No log 3.3867 254 1.2518 0.5161 1.2518 1.1189
No log 3.4133 256 1.3334 0.4849 1.3334 1.1547
No log 3.44 258 1.2584 0.4572 1.2584 1.1218
No log 3.4667 260 1.0610 0.5531 1.0610 1.0300
No log 3.4933 262 0.9276 0.5846 0.9276 0.9631
No log 3.52 264 0.8956 0.6174 0.8956 0.9464
No log 3.5467 266 0.9001 0.6289 0.9001 0.9487
No log 3.5733 268 0.9164 0.6277 0.9164 0.9573
No log 3.6 270 0.8648 0.6389 0.8648 0.9299
No log 3.6267 272 0.8204 0.6361 0.8204 0.9058
No log 3.6533 274 0.7645 0.6497 0.7645 0.8744
No log 3.68 276 0.8236 0.6469 0.8236 0.9075
No log 3.7067 278 0.8406 0.6323 0.8406 0.9169
No log 3.7333 280 0.7794 0.6490 0.7794 0.8828
No log 3.76 282 0.7670 0.6569 0.7670 0.8758
No log 3.7867 284 0.7524 0.6436 0.7524 0.8674
No log 3.8133 286 0.7689 0.6569 0.7689 0.8768
No log 3.84 288 0.8633 0.6403 0.8633 0.9292
No log 3.8667 290 0.8798 0.6261 0.8798 0.9380
No log 3.8933 292 0.8515 0.6404 0.8515 0.9228
No log 3.92 294 0.9290 0.5962 0.9290 0.9638
No log 3.9467 296 0.8718 0.6189 0.8718 0.9337
No log 3.9733 298 0.7893 0.6032 0.7893 0.8884
No log 4.0 300 0.8534 0.6030 0.8534 0.9238
No log 4.0267 302 1.0810 0.5588 1.0810 1.0397
No log 4.0533 304 1.3711 0.4951 1.3711 1.1709
No log 4.08 306 1.3995 0.5168 1.3995 1.1830
No log 4.1067 308 1.2414 0.5501 1.2414 1.1142
No log 4.1333 310 1.0356 0.5846 1.0356 1.0177
No log 4.16 312 0.9439 0.6074 0.9439 0.9716
No log 4.1867 314 0.8618 0.6270 0.8618 0.9283
No log 4.2133 316 0.8870 0.5947 0.8870 0.9418
No log 4.24 318 0.9164 0.6005 0.9164 0.9573
No log 4.2667 320 0.9941 0.6001 0.9941 0.9970
No log 4.2933 322 1.0835 0.5650 1.0835 1.0409
No log 4.32 324 1.0021 0.5832 1.0021 1.0011
No log 4.3467 326 0.8902 0.6349 0.8902 0.9435
No log 4.3733 328 0.9027 0.6393 0.9027 0.9501
No log 4.4 330 0.9960 0.5878 0.9960 0.9980
No log 4.4267 332 1.1782 0.5146 1.1782 1.0855
No log 4.4533 334 1.3183 0.4441 1.3183 1.1482
No log 4.48 336 1.2951 0.4225 1.2951 1.1380
No log 4.5067 338 1.0863 0.5205 1.0863 1.0423
No log 4.5333 340 1.0105 0.5523 1.0105 1.0052
No log 4.5600 342 1.0850 0.5284 1.0850 1.0416
No log 4.5867 344 1.2811 0.4543 1.2811 1.1319
No log 4.6133 346 1.2967 0.4454 1.2967 1.1387
No log 4.64 348 1.0732 0.5374 1.0732 1.0359
No log 4.6667 350 0.8207 0.5851 0.8207 0.9059
No log 4.6933 352 0.7004 0.5791 0.7004 0.8369
No log 4.72 354 0.6911 0.6155 0.6911 0.8313
No log 4.7467 356 0.7121 0.6472 0.7121 0.8439
No log 4.7733 358 0.9224 0.6070 0.9224 0.9604
No log 4.8 360 1.2460 0.4679 1.2460 1.1163
No log 4.8267 362 1.2919 0.4642 1.2919 1.1366
No log 4.8533 364 1.0682 0.5731 1.0682 1.0335
No log 4.88 366 0.8106 0.6304 0.8106 0.9003
No log 4.9067 368 0.7174 0.6581 0.7174 0.8470
No log 4.9333 370 0.7304 0.6530 0.7304 0.8546
No log 4.96 372 0.8112 0.6265 0.8112 0.9007
No log 4.9867 374 0.9258 0.5945 0.9258 0.9622
No log 5.0133 376 0.9809 0.6041 0.9809 0.9904
No log 5.04 378 0.9474 0.5941 0.9474 0.9733
No log 5.0667 380 0.8047 0.6464 0.8047 0.8970
No log 5.0933 382 0.7172 0.6643 0.7172 0.8469
No log 5.12 384 0.7368 0.6614 0.7368 0.8584
No log 5.1467 386 0.8447 0.6315 0.8447 0.9191
No log 5.1733 388 1.0599 0.5888 1.0599 1.0295
No log 5.2 390 1.2534 0.5102 1.2534 1.1196
No log 5.2267 392 1.2701 0.5008 1.2701 1.1270
No log 5.2533 394 1.1538 0.5175 1.1538 1.0741
No log 5.28 396 1.0092 0.5681 1.0092 1.0046
No log 5.3067 398 0.8586 0.6169 0.8586 0.9266
No log 5.3333 400 0.8102 0.6276 0.8102 0.9001
No log 5.36 402 0.8269 0.6394 0.8269 0.9093
No log 5.3867 404 0.9117 0.5994 0.9117 0.9548
No log 5.4133 406 0.9958 0.5580 0.9958 0.9979
No log 5.44 408 0.9982 0.5660 0.9982 0.9991
No log 5.4667 410 0.8980 0.6202 0.8980 0.9476
No log 5.4933 412 0.8319 0.6220 0.8319 0.9121
No log 5.52 414 0.7971 0.6421 0.7971 0.8928
No log 5.5467 416 0.7992 0.6439 0.7992 0.8940
No log 5.5733 418 0.7828 0.6551 0.7828 0.8848
No log 5.6 420 0.7893 0.6532 0.7893 0.8884
No log 5.6267 422 0.8988 0.6134 0.8988 0.9481
No log 5.6533 424 1.1017 0.5450 1.1017 1.0496
No log 5.68 426 1.2792 0.4894 1.2792 1.1310
No log 5.7067 428 1.2680 0.4894 1.2680 1.1261
No log 5.7333 430 1.1446 0.5248 1.1446 1.0699
No log 5.76 432 0.9179 0.6021 0.9179 0.9581
No log 5.7867 434 0.7967 0.6416 0.7967 0.8926
No log 5.8133 436 0.7534 0.6454 0.7534 0.8680
No log 5.84 438 0.7598 0.6485 0.7598 0.8717
No log 5.8667 440 0.8024 0.6560 0.8024 0.8958
No log 5.8933 442 0.7978 0.6717 0.7978 0.8932
No log 5.92 444 0.8138 0.6546 0.8138 0.9021
No log 5.9467 446 0.7807 0.6615 0.7807 0.8836
No log 5.9733 448 0.7264 0.6684 0.7264 0.8523
No log 6.0 450 0.6941 0.6764 0.6941 0.8331
No log 6.0267 452 0.7124 0.6572 0.7124 0.8440
No log 6.0533 454 0.7774 0.6635 0.7774 0.8817
No log 6.08 456 0.8527 0.6245 0.8527 0.9234
No log 6.1067 458 0.8691 0.6400 0.8691 0.9322
No log 6.1333 460 0.8180 0.6511 0.8180 0.9044
No log 6.16 462 0.7237 0.6491 0.7237 0.8507
No log 6.1867 464 0.6972 0.6308 0.6972 0.8350
No log 6.2133 466 0.6912 0.6443 0.6912 0.8314
No log 6.24 468 0.6915 0.6516 0.6915 0.8316
No log 6.2667 470 0.7880 0.6475 0.7880 0.8877
No log 6.2933 472 0.9439 0.6050 0.9439 0.9715
No log 6.32 474 1.0302 0.5741 1.0302 1.0150
No log 6.3467 476 1.0606 0.5639 1.0606 1.0298
No log 6.3733 478 1.0053 0.5876 1.0053 1.0027
No log 6.4 480 0.9661 0.5855 0.9661 0.9829
No log 6.4267 482 0.8610 0.6344 0.8610 0.9279
No log 6.4533 484 0.7473 0.6514 0.7473 0.8645
No log 6.48 486 0.7202 0.6796 0.7202 0.8487
No log 6.5067 488 0.7922 0.6472 0.7922 0.8901
No log 6.5333 490 0.9014 0.6126 0.9014 0.9494
No log 6.5600 492 1.0186 0.6154 1.0186 1.0093
No log 6.5867 494 1.0998 0.5915 1.0998 1.0487
No log 6.6133 496 1.0368 0.6165 1.0368 1.0182
No log 6.64 498 0.8878 0.6413 0.8878 0.9422
0.489 6.6667 500 0.7843 0.6539 0.7843 0.8856
0.489 6.6933 502 0.7839 0.6477 0.7839 0.8854
0.489 6.72 504 0.8685 0.6293 0.8685 0.9320
0.489 6.7467 506 0.9048 0.6388 0.9048 0.9512
0.489 6.7733 508 0.8514 0.6239 0.8514 0.9227
0.489 6.8 510 0.8154 0.6517 0.8154 0.9030

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k15_task1_organization

Finetuned
(4023)
this model