ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k3_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8351
  • Qwk: 0.3105
  • Mse: 0.8351
  • Rmse: 0.9138

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.1333 2 2.5493 0.0231 2.5493 1.5967
No log 0.2667 4 1.3249 0.1710 1.3249 1.1511
No log 0.4 6 0.7284 0.0937 0.7284 0.8534
No log 0.5333 8 1.0153 -0.0630 1.0153 1.0076
No log 0.6667 10 1.1062 -0.0186 1.1062 1.0518
No log 0.8 12 1.2104 0.0450 1.2104 1.1002
No log 0.9333 14 1.1081 0.1148 1.1081 1.0527
No log 1.0667 16 0.9574 0.3409 0.9574 0.9785
No log 1.2 18 0.8839 0.3498 0.8839 0.9402
No log 1.3333 20 0.8642 0.3224 0.8642 0.9296
No log 1.4667 22 0.8398 0.2261 0.8398 0.9164
No log 1.6 24 0.8220 0.1648 0.8220 0.9067
No log 1.7333 26 0.7717 0.3789 0.7717 0.8785
No log 1.8667 28 0.7504 0.3494 0.7504 0.8663
No log 2.0 30 0.7380 0.3564 0.7380 0.8591
No log 2.1333 32 0.7728 0.3494 0.7728 0.8791
No log 2.2667 34 0.7441 0.3789 0.7441 0.8626
No log 2.4 36 0.7001 0.3387 0.7001 0.8367
No log 2.5333 38 0.7386 0.3789 0.7386 0.8594
No log 2.6667 40 0.8248 0.3731 0.8248 0.9082
No log 2.8 42 0.8716 0.4214 0.8716 0.9336
No log 2.9333 44 0.8749 0.3890 0.8749 0.9354
No log 3.0667 46 0.8326 0.3913 0.8326 0.9125
No log 3.2 48 0.7619 0.2244 0.7619 0.8729
No log 3.3333 50 0.6595 0.3032 0.6595 0.8121
No log 3.4667 52 0.6288 0.3940 0.6288 0.7929
No log 3.6 54 0.5833 0.3918 0.5833 0.7637
No log 3.7333 56 0.5740 0.3755 0.5740 0.7577
No log 3.8667 58 0.5904 0.4614 0.5904 0.7684
No log 4.0 60 0.6626 0.4513 0.6626 0.8140
No log 4.1333 62 0.9079 0.3909 0.9079 0.9528
No log 4.2667 64 1.3102 0.2627 1.3102 1.1447
No log 4.4 66 1.3503 0.2752 1.3503 1.1620
No log 4.5333 68 1.0563 0.3730 1.0563 1.0278
No log 4.6667 70 0.7269 0.4190 0.7269 0.8526
No log 4.8 72 0.6697 0.3782 0.6697 0.8184
No log 4.9333 74 0.6788 0.4167 0.6788 0.8239
No log 5.0667 76 0.7779 0.2964 0.7779 0.8820
No log 5.2 78 0.9147 0.3370 0.9147 0.9564
No log 5.3333 80 1.1108 0.3412 1.1108 1.0539
No log 5.4667 82 1.1159 0.2968 1.1159 1.0563
No log 5.6 84 1.0737 0.3013 1.0737 1.0362
No log 5.7333 86 0.8418 0.4142 0.8418 0.9175
No log 5.8667 88 0.6210 0.4644 0.6210 0.7880
No log 6.0 90 0.5582 0.5228 0.5582 0.7472
No log 6.1333 92 0.5604 0.4837 0.5604 0.7486
No log 6.2667 94 0.5788 0.4322 0.5788 0.7608
No log 6.4 96 0.6349 0.4413 0.6349 0.7968
No log 6.5333 98 0.8074 0.3824 0.8074 0.8986
No log 6.6667 100 0.9658 0.3650 0.9658 0.9827
No log 6.8 102 1.0008 0.3849 1.0008 1.0004
No log 6.9333 104 0.8691 0.4032 0.8691 0.9322
No log 7.0667 106 0.7215 0.3770 0.7215 0.8494
No log 7.2 108 0.6547 0.3399 0.6547 0.8092
No log 7.3333 110 0.6465 0.4247 0.6465 0.8040
No log 7.4667 112 0.7029 0.4203 0.7029 0.8384
No log 7.6 114 0.7912 0.3719 0.7912 0.8895
No log 7.7333 116 0.8496 0.3868 0.8496 0.9218
No log 7.8667 118 0.8036 0.4021 0.8036 0.8964
No log 8.0 120 0.7125 0.5173 0.7125 0.8441
No log 8.1333 122 0.6024 0.5292 0.6024 0.7761
No log 8.2667 124 0.5942 0.5158 0.5942 0.7708
No log 8.4 126 0.6329 0.5201 0.6329 0.7955
No log 8.5333 128 0.8065 0.3868 0.8065 0.8981
No log 8.6667 130 1.0598 0.3999 1.0598 1.0295
No log 8.8 132 0.9645 0.3317 0.9645 0.9821
No log 8.9333 134 0.8018 0.3782 0.8018 0.8955
No log 9.0667 136 0.7435 0.3913 0.7435 0.8622
No log 9.2 138 0.8662 0.3707 0.8662 0.9307
No log 9.3333 140 0.9891 0.3193 0.9891 0.9945
No log 9.4667 142 0.9330 0.3716 0.9330 0.9659
No log 9.6 144 0.7595 0.4910 0.7595 0.8715
No log 9.7333 146 0.5726 0.5468 0.5726 0.7567
No log 9.8667 148 0.5413 0.4743 0.5413 0.7357
No log 10.0 150 0.5982 0.5678 0.5982 0.7734
No log 10.1333 152 0.8166 0.4900 0.8166 0.9037
No log 10.2667 154 0.9982 0.3915 0.9982 0.9991
No log 10.4 156 0.9698 0.3617 0.9698 0.9848
No log 10.5333 158 0.7829 0.3905 0.7829 0.8848
No log 10.6667 160 0.6588 0.3799 0.6588 0.8117
No log 10.8 162 0.6963 0.3433 0.6963 0.8344
No log 10.9333 164 0.8707 0.3556 0.8707 0.9331
No log 11.0667 166 1.0240 0.3690 1.0240 1.0119
No log 11.2 168 0.8801 0.3523 0.8801 0.9381
No log 11.3333 170 0.7586 0.3767 0.7586 0.8710
No log 11.4667 172 0.6800 0.3913 0.6800 0.8246
No log 11.6 174 0.6034 0.4123 0.6034 0.7768
No log 11.7333 176 0.6285 0.4295 0.6285 0.7928
No log 11.8667 178 0.6631 0.4018 0.6631 0.8143
No log 12.0 180 0.6351 0.4197 0.6351 0.7969
No log 12.1333 182 0.5719 0.4925 0.5719 0.7562
No log 12.2667 184 0.5400 0.5271 0.5400 0.7348
No log 12.4 186 0.5628 0.4749 0.5628 0.7502
No log 12.5333 188 0.5851 0.4409 0.5851 0.7649
No log 12.6667 190 0.6748 0.4275 0.6748 0.8215
No log 12.8 192 0.8672 0.4270 0.8672 0.9312
No log 12.9333 194 1.0545 0.3473 1.0545 1.0269
No log 13.0667 196 1.0958 0.3367 1.0958 1.0468
No log 13.2 198 0.9996 0.3390 0.9996 0.9998
No log 13.3333 200 0.8276 0.3477 0.8276 0.9097
No log 13.4667 202 0.7088 0.3819 0.7088 0.8419
No log 13.6 204 0.6725 0.3950 0.6725 0.8201
No log 13.7333 206 0.6679 0.4167 0.6679 0.8172
No log 13.8667 208 0.7575 0.3731 0.7575 0.8703
No log 14.0 210 0.9587 0.3850 0.9587 0.9791
No log 14.1333 212 1.0213 0.3798 1.0213 1.0106
No log 14.2667 214 0.9915 0.3850 0.9915 0.9957
No log 14.4 216 0.8272 0.4081 0.8272 0.9095
No log 14.5333 218 0.7287 0.4333 0.7287 0.8536
No log 14.6667 220 0.7101 0.4400 0.7101 0.8427
No log 14.8 222 0.7448 0.4684 0.7448 0.8630
No log 14.9333 224 0.7801 0.3906 0.7801 0.8832
No log 15.0667 226 0.8719 0.3956 0.8719 0.9337
No log 15.2 228 0.9016 0.3776 0.9016 0.9495
No log 15.3333 230 0.9649 0.3812 0.9649 0.9823
No log 15.4667 232 0.8951 0.4166 0.8951 0.9461
No log 15.6 234 0.8792 0.4081 0.8792 0.9377
No log 15.7333 236 0.9782 0.3655 0.9782 0.9890
No log 15.8667 238 0.9815 0.3622 0.9815 0.9907
No log 16.0 240 0.8973 0.3830 0.8973 0.9472
No log 16.1333 242 0.7998 0.4381 0.7998 0.8943
No log 16.2667 244 0.7469 0.4745 0.7469 0.8642
No log 16.4 246 0.7527 0.4424 0.7527 0.8676
No log 16.5333 248 0.8494 0.3499 0.8494 0.9216
No log 16.6667 250 0.9253 0.3269 0.9253 0.9619
No log 16.8 252 0.9764 0.3324 0.9764 0.9881
No log 16.9333 254 0.8790 0.3538 0.8790 0.9375
No log 17.0667 256 0.8225 0.3601 0.8225 0.9069
No log 17.2 258 0.8044 0.3665 0.8044 0.8969
No log 17.3333 260 0.8509 0.3251 0.8509 0.9225
No log 17.4667 262 0.8667 0.3193 0.8667 0.9310
No log 17.6 264 0.8372 0.3194 0.8372 0.9150
No log 17.7333 266 0.8507 0.3169 0.8507 0.9224
No log 17.8667 268 0.8443 0.3799 0.8443 0.9189
No log 18.0 270 0.8630 0.3193 0.8630 0.9290
No log 18.1333 272 0.8865 0.3381 0.8865 0.9415
No log 18.2667 274 0.8479 0.4333 0.8479 0.9208
No log 18.4 276 0.8899 0.4488 0.8899 0.9434
No log 18.5333 278 0.9341 0.4437 0.9341 0.9665
No log 18.6667 280 0.9112 0.4789 0.9112 0.9546
No log 18.8 282 0.9585 0.4015 0.9585 0.9790
No log 18.9333 284 0.9283 0.4015 0.9283 0.9635
No log 19.0667 286 0.8362 0.4189 0.8362 0.9144
No log 19.2 288 0.7110 0.4587 0.7110 0.8432
No log 19.3333 290 0.6823 0.4827 0.6823 0.8260
No log 19.4667 292 0.7431 0.4512 0.7431 0.8620
No log 19.6 294 0.8076 0.3559 0.8076 0.8987
No log 19.7333 296 0.7516 0.4502 0.7516 0.8669
No log 19.8667 298 0.7681 0.4587 0.7681 0.8764
No log 20.0 300 0.8079 0.3719 0.8079 0.8988
No log 20.1333 302 0.8371 0.3868 0.8371 0.9149
No log 20.2667 304 0.8737 0.3608 0.8737 0.9347
No log 20.4 306 0.7678 0.4081 0.7678 0.8762
No log 20.5333 308 0.6922 0.4812 0.6922 0.8320
No log 20.6667 310 0.5839 0.5586 0.5839 0.7642
No log 20.8 312 0.5687 0.5104 0.5687 0.7541
No log 20.9333 314 0.6186 0.5081 0.6186 0.7865
No log 21.0667 316 0.7653 0.5416 0.7653 0.8748
No log 21.2 318 1.0082 0.4118 1.0082 1.0041
No log 21.3333 320 0.9770 0.4155 0.9770 0.9884
No log 21.4667 322 0.7903 0.3953 0.7903 0.8890
No log 21.6 324 0.6298 0.5219 0.6298 0.7936
No log 21.7333 326 0.6027 0.4855 0.6027 0.7763
No log 21.8667 328 0.6441 0.5067 0.6441 0.8025
No log 22.0 330 0.7546 0.4801 0.7546 0.8687
No log 22.1333 332 0.7640 0.4726 0.7640 0.8741
No log 22.2667 334 0.8007 0.4646 0.8007 0.8948
No log 22.4 336 0.7355 0.4808 0.7355 0.8576
No log 22.5333 338 0.7104 0.4892 0.7104 0.8429
No log 22.6667 340 0.6976 0.4892 0.6976 0.8352
No log 22.8 342 0.7113 0.5131 0.7113 0.8434
No log 22.9333 344 0.7079 0.5163 0.7079 0.8414
No log 23.0667 346 0.7352 0.5065 0.7352 0.8575
No log 23.2 348 0.7910 0.4668 0.7910 0.8894
No log 23.3333 350 0.7998 0.4794 0.7998 0.8943
No log 23.4667 352 0.7564 0.4867 0.7564 0.8697
No log 23.6 354 0.6923 0.5410 0.6923 0.8320
No log 23.7333 356 0.6844 0.5410 0.6844 0.8273
No log 23.8667 358 0.7565 0.4536 0.7565 0.8698
No log 24.0 360 0.8506 0.4189 0.8506 0.9223
No log 24.1333 362 0.9039 0.4023 0.9039 0.9507
No log 24.2667 364 0.9273 0.3697 0.9273 0.9629
No log 24.4 366 0.8155 0.4400 0.8155 0.9031
No log 24.5333 368 0.7084 0.5586 0.7084 0.8417
No log 24.6667 370 0.7079 0.5639 0.7079 0.8414
No log 24.8 372 0.8155 0.4333 0.8155 0.9031
No log 24.9333 374 0.9007 0.3523 0.9007 0.9490
No log 25.0667 376 0.9871 0.3681 0.9871 0.9935
No log 25.2 378 0.9324 0.3643 0.9324 0.9656
No log 25.3333 380 0.8241 0.3707 0.8241 0.9078
No log 25.4667 382 0.7190 0.4036 0.7190 0.8479
No log 25.6 384 0.6914 0.4112 0.6914 0.8315
No log 25.7333 386 0.7318 0.3981 0.7318 0.8555
No log 25.8667 388 0.8246 0.3807 0.8246 0.9081
No log 26.0 390 0.8901 0.3425 0.8901 0.9435
No log 26.1333 392 0.9412 0.3233 0.9412 0.9702
No log 26.2667 394 0.9149 0.3484 0.9149 0.9565
No log 26.4 396 0.8546 0.4114 0.8546 0.9245
No log 26.5333 398 0.7985 0.4531 0.7985 0.8936
No log 26.6667 400 0.7304 0.5086 0.7304 0.8546
No log 26.8 402 0.7419 0.4978 0.7419 0.8613
No log 26.9333 404 0.8044 0.3913 0.8044 0.8969
No log 27.0667 406 0.8549 0.3445 0.8549 0.9246
No log 27.2 408 0.8232 0.3499 0.8232 0.9073
No log 27.3333 410 0.7618 0.4349 0.7618 0.8728
No log 27.4667 412 0.7169 0.4576 0.7169 0.8467
No log 27.6 414 0.6757 0.4925 0.6757 0.8220
No log 27.7333 416 0.6996 0.4911 0.6996 0.8364
No log 27.8667 418 0.7370 0.4819 0.7370 0.8585
No log 28.0 420 0.7276 0.4819 0.7276 0.8530
No log 28.1333 422 0.6672 0.5452 0.6672 0.8168
No log 28.2667 424 0.6516 0.5403 0.6516 0.8072
No log 28.4 426 0.6580 0.5349 0.6580 0.8112
No log 28.5333 428 0.7039 0.4741 0.7039 0.8390
No log 28.6667 430 0.8036 0.3669 0.8036 0.8964
No log 28.8 432 0.8440 0.3669 0.8440 0.9187
No log 28.9333 434 0.7702 0.3953 0.7702 0.8776
No log 29.0667 436 0.7142 0.4369 0.7142 0.8451
No log 29.2 438 0.6378 0.6029 0.6378 0.7986
No log 29.3333 440 0.6170 0.6029 0.6170 0.7855
No log 29.4667 442 0.6596 0.4602 0.6596 0.8122
No log 29.6 444 0.6906 0.4350 0.6906 0.8310
No log 29.7333 446 0.6625 0.4350 0.6625 0.8140
No log 29.8667 448 0.6125 0.6029 0.6125 0.7826
No log 30.0 450 0.5694 0.5803 0.5694 0.7546
No log 30.1333 452 0.5637 0.5803 0.5637 0.7508
No log 30.2667 454 0.5816 0.5803 0.5816 0.7626
No log 30.4 456 0.6320 0.6029 0.6320 0.7950
No log 30.5333 458 0.7291 0.4606 0.7291 0.8539
No log 30.6667 460 0.8245 0.3577 0.8245 0.9080
No log 30.8 462 0.7912 0.3909 0.7912 0.8895
No log 30.9333 464 0.7132 0.4522 0.7132 0.8445
No log 31.0667 466 0.6629 0.5252 0.6629 0.8142
No log 31.2 468 0.6548 0.5252 0.6548 0.8092
No log 31.3333 470 0.6921 0.5310 0.6921 0.8319
No log 31.4667 472 0.7014 0.5544 0.7014 0.8375
No log 31.6 474 0.7026 0.5586 0.7026 0.8382
No log 31.7333 476 0.7362 0.4457 0.7362 0.8580
No log 31.8667 478 0.7525 0.4385 0.7525 0.8675
No log 32.0 480 0.7509 0.4385 0.7509 0.8665
No log 32.1333 482 0.7135 0.5190 0.7135 0.8447
No log 32.2667 484 0.6821 0.4997 0.6821 0.8259
No log 32.4 486 0.7247 0.5219 0.7247 0.8513
No log 32.5333 488 0.8262 0.2971 0.8262 0.9090
No log 32.6667 490 0.8992 0.3131 0.8992 0.9483
No log 32.8 492 0.9191 0.2898 0.9191 0.9587
No log 32.9333 494 0.8656 0.3105 0.8656 0.9304
No log 33.0667 496 0.7626 0.3940 0.7626 0.8732
No log 33.2 498 0.7118 0.4997 0.7118 0.8437
0.2561 33.3333 500 0.7257 0.4491 0.7257 0.8519
0.2561 33.4667 502 0.7575 0.4329 0.7575 0.8704
0.2561 33.6 504 0.8212 0.3076 0.8212 0.9062
0.2561 33.7333 506 0.8982 0.2461 0.8982 0.9477
0.2561 33.8667 508 0.8868 0.2777 0.8868 0.9417
0.2561 34.0 510 0.8351 0.3105 0.8351 0.9138

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k3_task7_organization

Finetuned
(4019)
this model