ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k7_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5632
  • Qwk: 0.5774
  • Mse: 0.5632
  • Rmse: 0.7505

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0541 2 3.9749 -0.0254 3.9749 1.9937
No log 0.1081 4 2.3789 -0.0086 2.3789 1.5424
No log 0.1622 6 2.3060 -0.0206 2.3060 1.5185
No log 0.2162 8 1.8982 -0.0169 1.8982 1.3777
No log 0.2703 10 1.1726 0.2687 1.1726 1.0829
No log 0.3243 12 1.0136 0.1881 1.0136 1.0068
No log 0.3784 14 0.9781 0.1292 0.9781 0.9890
No log 0.4324 16 0.9785 0.1981 0.9785 0.9892
No log 0.4865 18 1.1178 0.1821 1.1178 1.0573
No log 0.5405 20 1.3287 -0.0328 1.3287 1.1527
No log 0.5946 22 1.2122 0.0938 1.2122 1.1010
No log 0.6486 24 1.0211 0.2958 1.0211 1.0105
No log 0.7027 26 0.9265 0.3498 0.9265 0.9625
No log 0.7568 28 0.9039 0.4090 0.9039 0.9507
No log 0.8108 30 1.0448 0.3229 1.0448 1.0222
No log 0.8649 32 1.2718 -0.0578 1.2718 1.1277
No log 0.9189 34 1.1846 0.1062 1.1846 1.0884
No log 0.9730 36 1.1816 0.0462 1.1816 1.0870
No log 1.0270 38 1.2050 0.0584 1.2050 1.0977
No log 1.0811 40 1.0468 0.2318 1.0468 1.0231
No log 1.1351 42 0.8819 0.3646 0.8819 0.9391
No log 1.1892 44 0.8572 0.3266 0.8572 0.9258
No log 1.2432 46 0.9094 0.4180 0.9094 0.9536
No log 1.2973 48 0.8930 0.3221 0.8930 0.9450
No log 1.3514 50 0.8941 0.3506 0.8941 0.9456
No log 1.4054 52 0.8985 0.3214 0.8985 0.9479
No log 1.4595 54 0.9153 0.2967 0.9153 0.9567
No log 1.5135 56 1.0002 0.2588 1.0002 1.0001
No log 1.5676 58 0.9516 0.2880 0.9516 0.9755
No log 1.6216 60 0.8100 0.4692 0.8100 0.9000
No log 1.6757 62 0.7851 0.3817 0.7851 0.8861
No log 1.7297 64 0.8001 0.3797 0.8001 0.8945
No log 1.7838 66 0.7779 0.4373 0.7779 0.8820
No log 1.8378 68 0.7866 0.5093 0.7866 0.8869
No log 1.8919 70 0.7695 0.4944 0.7695 0.8772
No log 1.9459 72 0.8222 0.4728 0.8222 0.9067
No log 2.0 74 1.0264 0.3595 1.0264 1.0131
No log 2.0541 76 1.0866 0.3506 1.0866 1.0424
No log 2.1081 78 0.8823 0.3744 0.8823 0.9393
No log 2.1622 80 0.7437 0.4540 0.7437 0.8624
No log 2.2162 82 0.7611 0.4708 0.7611 0.8724
No log 2.2703 84 0.7523 0.4708 0.7523 0.8674
No log 2.3243 86 0.7584 0.4960 0.7584 0.8708
No log 2.3784 88 0.7062 0.4540 0.7062 0.8404
No log 2.4324 90 0.7178 0.5721 0.7178 0.8472
No log 2.4865 92 0.7166 0.5627 0.7166 0.8465
No log 2.5405 94 0.7288 0.5025 0.7288 0.8537
No log 2.5946 96 0.7582 0.4156 0.7582 0.8708
No log 2.6486 98 0.7673 0.4444 0.7673 0.8760
No log 2.7027 100 0.7561 0.4407 0.7561 0.8695
No log 2.7568 102 0.8624 0.5154 0.8624 0.9286
No log 2.8108 104 1.0346 0.3902 1.0346 1.0171
No log 2.8649 106 0.9928 0.4524 0.9928 0.9964
No log 2.9189 108 0.8232 0.4826 0.8232 0.9073
No log 2.9730 110 0.7940 0.4761 0.7940 0.8911
No log 3.0270 112 0.8212 0.5024 0.8212 0.9062
No log 3.0811 114 0.8411 0.4600 0.8411 0.9171
No log 3.1351 116 0.7773 0.5572 0.7773 0.8817
No log 3.1892 118 0.7042 0.5558 0.7042 0.8392
No log 3.2432 120 0.6855 0.5221 0.6855 0.8280
No log 3.2973 122 0.6878 0.5208 0.6878 0.8293
No log 3.3514 124 0.7430 0.5181 0.7430 0.8620
No log 3.4054 126 0.7902 0.4666 0.7902 0.8890
No log 3.4595 128 0.6883 0.5220 0.6883 0.8297
No log 3.5135 130 0.6954 0.5287 0.6954 0.8339
No log 3.5676 132 0.6808 0.5505 0.6808 0.8251
No log 3.6216 134 0.6600 0.6065 0.6600 0.8124
No log 3.6757 136 0.6703 0.5933 0.6703 0.8187
No log 3.7297 138 0.6981 0.5707 0.6981 0.8355
No log 3.7838 140 0.6528 0.6219 0.6528 0.8080
No log 3.8378 142 0.6592 0.5590 0.6592 0.8119
No log 3.8919 144 0.6708 0.5590 0.6708 0.8190
No log 3.9459 146 0.6894 0.5146 0.6894 0.8303
No log 4.0 148 0.7196 0.4792 0.7196 0.8483
No log 4.0541 150 0.7279 0.5678 0.7279 0.8532
No log 4.1081 152 0.7576 0.4975 0.7576 0.8704
No log 4.1622 154 0.7516 0.5558 0.7516 0.8670
No log 4.2162 156 0.7352 0.5700 0.7352 0.8574
No log 4.2703 158 0.7192 0.5905 0.7192 0.8481
No log 4.3243 160 0.7060 0.5185 0.7060 0.8402
No log 4.3784 162 0.7166 0.4841 0.7166 0.8465
No log 4.4324 164 0.6657 0.5048 0.6657 0.8159
No log 4.4865 166 0.6845 0.4608 0.6845 0.8273
No log 4.5405 168 0.6785 0.4867 0.6785 0.8237
No log 4.5946 170 0.6883 0.5292 0.6883 0.8296
No log 4.6486 172 0.7065 0.4812 0.7065 0.8405
No log 4.7027 174 0.7008 0.5555 0.7008 0.8371
No log 4.7568 176 0.6345 0.5844 0.6345 0.7966
No log 4.8108 178 0.6227 0.6187 0.6227 0.7891
No log 4.8649 180 0.6597 0.5959 0.6597 0.8122
No log 4.9189 182 0.8235 0.4976 0.8235 0.9075
No log 4.9730 184 0.9899 0.5462 0.9899 0.9950
No log 5.0270 186 0.8608 0.4667 0.8608 0.9278
No log 5.0811 188 0.6907 0.5170 0.6907 0.8311
No log 5.1351 190 0.6263 0.5943 0.6263 0.7914
No log 5.1892 192 0.6457 0.6634 0.6457 0.8036
No log 5.2432 194 0.6887 0.6605 0.6887 0.8299
No log 5.2973 196 0.7879 0.5621 0.7879 0.8876
No log 5.3514 198 0.8848 0.4693 0.8848 0.9406
No log 5.4054 200 0.7490 0.5988 0.7490 0.8655
No log 5.4595 202 0.6376 0.5923 0.6376 0.7985
No log 5.5135 204 0.6968 0.5993 0.6968 0.8347
No log 5.5676 206 0.6784 0.5782 0.6784 0.8237
No log 5.6216 208 0.6573 0.6133 0.6573 0.8107
No log 5.6757 210 0.7490 0.5046 0.7490 0.8654
No log 5.7297 212 0.7855 0.5041 0.7855 0.8863
No log 5.7838 214 0.7334 0.5084 0.7334 0.8564
No log 5.8378 216 0.7038 0.4511 0.7038 0.8389
No log 5.8919 218 0.7052 0.5302 0.7052 0.8398
No log 5.9459 220 0.7050 0.5529 0.7050 0.8396
No log 6.0 222 0.7002 0.4661 0.7002 0.8368
No log 6.0541 224 0.7336 0.4975 0.7336 0.8565
No log 6.1081 226 0.7545 0.5156 0.7545 0.8686
No log 6.1622 228 0.7151 0.5279 0.7151 0.8457
No log 6.2162 230 0.6747 0.4368 0.6747 0.8214
No log 6.2703 232 0.6602 0.5556 0.6602 0.8125
No log 6.3243 234 0.6598 0.5690 0.6598 0.8123
No log 6.3784 236 0.6614 0.5142 0.6614 0.8133
No log 6.4324 238 0.7064 0.4581 0.7064 0.8405
No log 6.4865 240 0.7210 0.4584 0.7210 0.8491
No log 6.5405 242 0.6926 0.4711 0.6926 0.8322
No log 6.5946 244 0.6739 0.5197 0.6739 0.8209
No log 6.6486 246 0.6722 0.5121 0.6722 0.8199
No log 6.7027 248 0.6985 0.6373 0.6985 0.8358
No log 6.7568 250 0.7729 0.5650 0.7729 0.8792
No log 6.8108 252 0.7129 0.6184 0.7129 0.8444
No log 6.8649 254 0.6568 0.5752 0.6568 0.8105
No log 6.9189 256 0.6856 0.5500 0.6856 0.8280
No log 6.9730 258 0.7020 0.5090 0.7020 0.8379
No log 7.0270 260 0.7415 0.4876 0.7415 0.8611
No log 7.0811 262 0.7363 0.5069 0.7363 0.8581
No log 7.1351 264 0.7260 0.4953 0.7260 0.8520
No log 7.1892 266 0.6998 0.5788 0.6998 0.8365
No log 7.2432 268 0.6316 0.5959 0.6316 0.7947
No log 7.2973 270 0.5980 0.5842 0.5980 0.7733
No log 7.3514 272 0.6102 0.5842 0.6102 0.7812
No log 7.4054 274 0.6860 0.5788 0.6860 0.8282
No log 7.4595 276 0.7719 0.5501 0.7719 0.8786
No log 7.5135 278 0.8192 0.5577 0.8192 0.9051
No log 7.5676 280 0.9811 0.5261 0.9811 0.9905
No log 7.6216 282 1.0112 0.5283 1.0112 1.0056
No log 7.6757 284 0.9006 0.5098 0.9006 0.9490
No log 7.7297 286 0.7731 0.5842 0.7731 0.8793
No log 7.7838 288 0.7109 0.5885 0.7109 0.8432
No log 7.8378 290 0.7276 0.5888 0.7276 0.8530
No log 7.8919 292 0.6901 0.5888 0.6901 0.8308
No log 7.9459 294 0.6468 0.5798 0.6468 0.8042
No log 8.0 296 0.6145 0.6547 0.6145 0.7839
No log 8.0541 298 0.6675 0.6184 0.6675 0.8170
No log 8.1081 300 0.7182 0.5837 0.7182 0.8475
No log 8.1622 302 0.7059 0.6174 0.7059 0.8402
No log 8.2162 304 0.6921 0.4975 0.6921 0.8319
No log 8.2703 306 0.7473 0.4826 0.7473 0.8645
No log 8.3243 308 0.7212 0.4725 0.7212 0.8492
No log 8.3784 310 0.7004 0.4995 0.7004 0.8369
No log 8.4324 312 0.6951 0.5210 0.6951 0.8337
No log 8.4865 314 0.7388 0.4820 0.7388 0.8595
No log 8.5405 316 0.7140 0.5039 0.7140 0.8450
No log 8.5946 318 0.6822 0.5516 0.6822 0.8259
No log 8.6486 320 0.6924 0.5383 0.6924 0.8321
No log 8.7027 322 0.7499 0.5032 0.7499 0.8659
No log 8.7568 324 0.7841 0.4581 0.7841 0.8855
No log 8.8108 326 0.9165 0.4869 0.9165 0.9573
No log 8.8649 328 0.9213 0.4769 0.9213 0.9598
No log 8.9189 330 0.8215 0.5020 0.8215 0.9064
No log 8.9730 332 0.7809 0.4916 0.7809 0.8837
No log 9.0270 334 0.6859 0.4818 0.6859 0.8282
No log 9.0811 336 0.6175 0.6003 0.6175 0.7858
No log 9.1351 338 0.6139 0.5726 0.6139 0.7835
No log 9.1892 340 0.6292 0.5350 0.6292 0.7932
No log 9.2432 342 0.6154 0.6185 0.6154 0.7845
No log 9.2973 344 0.6523 0.4940 0.6523 0.8077
No log 9.3514 346 0.7411 0.4921 0.7411 0.8608
No log 9.4054 348 0.7717 0.4911 0.7717 0.8785
No log 9.4595 350 0.7105 0.5026 0.7105 0.8429
No log 9.5135 352 0.6449 0.6473 0.6449 0.8030
No log 9.5676 354 0.6240 0.6838 0.6240 0.7899
No log 9.6216 356 0.6388 0.5869 0.6388 0.7992
No log 9.6757 358 0.7152 0.4921 0.7152 0.8457
No log 9.7297 360 0.7273 0.4696 0.7273 0.8528
No log 9.7838 362 0.6767 0.4832 0.6767 0.8226
No log 9.8378 364 0.6189 0.5546 0.6189 0.7867
No log 9.8919 366 0.5973 0.5894 0.5973 0.7729
No log 9.9459 368 0.5956 0.5894 0.5956 0.7717
No log 10.0 370 0.6323 0.5410 0.6323 0.7952
No log 10.0541 372 0.6474 0.5498 0.6474 0.8046
No log 10.1081 374 0.6206 0.6656 0.6206 0.7878
No log 10.1622 376 0.6039 0.6932 0.6039 0.7771
No log 10.2162 378 0.5928 0.6940 0.5928 0.7699
No log 10.2703 380 0.5781 0.6932 0.5781 0.7603
No log 10.3243 382 0.5702 0.6866 0.5702 0.7551
No log 10.3784 384 0.5866 0.6301 0.5866 0.7659
No log 10.4324 386 0.5850 0.6102 0.5850 0.7649
No log 10.4865 388 0.5623 0.6805 0.5623 0.7499
No log 10.5405 390 0.5695 0.6795 0.5695 0.7547
No log 10.5946 392 0.5507 0.6750 0.5507 0.7421
No log 10.6486 394 0.5292 0.7362 0.5292 0.7275
No log 10.7027 396 0.5596 0.6973 0.5596 0.7481
No log 10.7568 398 0.5713 0.6973 0.5713 0.7558
No log 10.8108 400 0.5541 0.7303 0.5541 0.7444
No log 10.8649 402 0.5594 0.6708 0.5594 0.7479
No log 10.9189 404 0.5631 0.6813 0.5631 0.7504
No log 10.9730 406 0.5636 0.6813 0.5636 0.7507
No log 11.0270 408 0.5760 0.6664 0.5760 0.7589
No log 11.0811 410 0.6040 0.5774 0.6040 0.7772
No log 11.1351 412 0.5967 0.5921 0.5967 0.7725
No log 11.1892 414 0.5973 0.6173 0.5973 0.7729
No log 11.2432 416 0.6006 0.6280 0.6006 0.7750
No log 11.2973 418 0.6338 0.4711 0.6338 0.7961
No log 11.3514 420 0.7428 0.5253 0.7428 0.8619
No log 11.4054 422 0.7465 0.5241 0.7465 0.8640
No log 11.4595 424 0.6444 0.4940 0.6444 0.8028
No log 11.5135 426 0.5814 0.6673 0.5814 0.7625
No log 11.5676 428 0.5859 0.7059 0.5859 0.7655
No log 11.6216 430 0.5996 0.6291 0.5996 0.7744
No log 11.6757 432 0.6197 0.5851 0.6197 0.7872
No log 11.7297 434 0.6280 0.5149 0.6280 0.7925
No log 11.7838 436 0.6479 0.4867 0.6479 0.8049
No log 11.8378 438 0.6414 0.4473 0.6414 0.8009
No log 11.8919 440 0.6184 0.6044 0.6184 0.7864
No log 11.9459 442 0.6007 0.6044 0.6007 0.7750
No log 12.0 444 0.5867 0.6380 0.5867 0.7659
No log 12.0541 446 0.5879 0.6405 0.5879 0.7668
No log 12.1081 448 0.5775 0.6510 0.5775 0.7599
No log 12.1622 450 0.5560 0.7089 0.5560 0.7457
No log 12.2162 452 0.5716 0.7154 0.5716 0.7560
No log 12.2703 454 0.5707 0.7533 0.5707 0.7555
No log 12.3243 456 0.5405 0.7245 0.5405 0.7352
No log 12.3784 458 0.5489 0.6916 0.5489 0.7409
No log 12.4324 460 0.5919 0.6167 0.5919 0.7694
No log 12.4865 462 0.6297 0.5833 0.6297 0.7935
No log 12.5405 464 0.6148 0.6167 0.6148 0.7841
No log 12.5946 466 0.6060 0.5869 0.6060 0.7785
No log 12.6486 468 0.6323 0.5975 0.6323 0.7952
No log 12.7027 470 0.6405 0.5975 0.6405 0.8003
No log 12.7568 472 0.5952 0.5975 0.5952 0.7715
No log 12.8108 474 0.5957 0.5975 0.5957 0.7718
No log 12.8649 476 0.6373 0.5975 0.6373 0.7983
No log 12.9189 478 0.7246 0.5219 0.7246 0.8513
No log 12.9730 480 0.6805 0.5372 0.6805 0.8250
No log 13.0270 482 0.5757 0.5975 0.5757 0.7588
No log 13.0811 484 0.5312 0.6879 0.5312 0.7288
No log 13.1351 486 0.5298 0.7118 0.5298 0.7279
No log 13.1892 488 0.5304 0.7042 0.5304 0.7283
No log 13.2432 490 0.5307 0.7176 0.5307 0.7285
No log 13.2973 492 0.5541 0.5774 0.5541 0.7444
No log 13.3514 494 0.6201 0.6377 0.6201 0.7875
No log 13.4054 496 0.6009 0.6377 0.6009 0.7752
No log 13.4595 498 0.5485 0.6241 0.5485 0.7406
0.3241 13.5135 500 0.5446 0.6648 0.5446 0.7380
0.3241 13.5676 502 0.5527 0.6456 0.5527 0.7435
0.3241 13.6216 504 0.5547 0.6736 0.5547 0.7448
0.3241 13.6757 506 0.5884 0.6328 0.5884 0.7671
0.3241 13.7297 508 0.6463 0.6004 0.6463 0.8040
0.3241 13.7838 510 0.6263 0.6109 0.6263 0.7914
0.3241 13.8378 512 0.5667 0.6328 0.5667 0.7528
0.3241 13.8919 514 0.5476 0.7081 0.5476 0.7400
0.3241 13.9459 516 0.5446 0.7124 0.5446 0.7380
0.3241 14.0 518 0.5531 0.6950 0.5531 0.7437
0.3241 14.0541 520 0.5666 0.5798 0.5666 0.7528
0.3241 14.1081 522 0.6093 0.6167 0.6093 0.7805
0.3241 14.1622 524 0.6472 0.5707 0.6472 0.8045
0.3241 14.2162 526 0.6225 0.5938 0.6225 0.7890
0.3241 14.2703 528 0.5632 0.5774 0.5632 0.7505

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k7_task5_organization

Finetuned
(4019)
this model