ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k7_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7143
  • Qwk: 0.4723
  • Mse: 0.7143
  • Rmse: 0.8451

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0541 2 2.5926 -0.0593 2.5926 1.6101
No log 0.1081 4 1.4086 0.0721 1.4086 1.1869
No log 0.1622 6 0.9673 -0.0970 0.9673 0.9835
No log 0.2162 8 0.6863 0.1633 0.6863 0.8285
No log 0.2703 10 0.6376 0.2843 0.6376 0.7985
No log 0.3243 12 0.6505 0.5083 0.6505 0.8065
No log 0.3784 14 1.0010 0.2207 1.0010 1.0005
No log 0.4324 16 0.7519 0.4922 0.7519 0.8671
No log 0.4865 18 0.6131 0.5841 0.6131 0.7830
No log 0.5405 20 0.5142 0.6156 0.5142 0.7170
No log 0.5946 22 0.5644 0.3942 0.5644 0.7513
No log 0.6486 24 0.5807 0.3372 0.5807 0.7621
No log 0.7027 26 0.5338 0.4238 0.5338 0.7306
No log 0.7568 28 0.5604 0.4677 0.5604 0.7486
No log 0.8108 30 0.5541 0.5222 0.5541 0.7444
No log 0.8649 32 0.5157 0.4044 0.5157 0.7181
No log 0.9189 34 0.6586 0.4482 0.6586 0.8115
No log 0.9730 36 0.6796 0.4482 0.6796 0.8244
No log 1.0270 38 0.6332 0.4568 0.6332 0.7957
No log 1.0811 40 0.5133 0.4722 0.5133 0.7164
No log 1.1351 42 0.5296 0.5414 0.5296 0.7277
No log 1.1892 44 0.5374 0.4958 0.5374 0.7331
No log 1.2432 46 0.5414 0.4701 0.5414 0.7358
No log 1.2973 48 0.5436 0.4322 0.5436 0.7373
No log 1.3514 50 0.5501 0.4019 0.5501 0.7417
No log 1.4054 52 0.6804 0.3777 0.6804 0.8248
No log 1.4595 54 0.7890 0.3777 0.7890 0.8883
No log 1.5135 56 0.8642 0.3538 0.8642 0.9296
No log 1.5676 58 0.6415 0.4089 0.6415 0.8009
No log 1.6216 60 0.5831 0.4659 0.5831 0.7636
No log 1.6757 62 0.5918 0.4596 0.5918 0.7693
No log 1.7297 64 0.7086 0.4243 0.7086 0.8418
No log 1.7838 66 0.7863 0.3867 0.7863 0.8868
No log 1.8378 68 0.7402 0.3867 0.7402 0.8604
No log 1.8919 70 0.6206 0.4954 0.6206 0.7878
No log 1.9459 72 0.6142 0.5164 0.6142 0.7837
No log 2.0 74 0.6046 0.4889 0.6046 0.7775
No log 2.0541 76 0.6450 0.4777 0.6450 0.8031
No log 2.1081 78 0.6675 0.4836 0.6675 0.8170
No log 2.1622 80 0.6190 0.4864 0.6190 0.7868
No log 2.2162 82 0.5859 0.5034 0.5859 0.7655
No log 2.2703 84 0.6104 0.4642 0.6104 0.7813
No log 2.3243 86 0.7062 0.3913 0.7062 0.8404
No log 2.3784 88 0.7241 0.3782 0.7241 0.8510
No log 2.4324 90 0.8163 0.4258 0.8163 0.9035
No log 2.4865 92 0.7084 0.4474 0.7084 0.8417
No log 2.5405 94 0.5330 0.5323 0.5330 0.7301
No log 2.5946 96 0.5395 0.5687 0.5395 0.7345
No log 2.6486 98 0.5469 0.5904 0.5469 0.7396
No log 2.7027 100 0.5232 0.5476 0.5232 0.7233
No log 2.7568 102 0.5260 0.5450 0.5260 0.7253
No log 2.8108 104 0.5858 0.5636 0.5858 0.7654
No log 2.8649 106 0.9177 0.4240 0.9177 0.9579
No log 2.9189 108 0.8737 0.4430 0.8737 0.9347
No log 2.9730 110 0.5681 0.5607 0.5681 0.7538
No log 3.0270 112 0.6510 0.4684 0.6510 0.8068
No log 3.0811 114 0.6547 0.4457 0.6547 0.8091
No log 3.1351 116 0.5462 0.4678 0.5462 0.7391
No log 3.1892 118 0.7471 0.4743 0.7471 0.8643
No log 3.2432 120 0.8099 0.4479 0.8099 0.8999
No log 3.2973 122 0.5906 0.4981 0.5906 0.7685
No log 3.3514 124 0.5881 0.4532 0.5881 0.7669
No log 3.4054 126 0.5967 0.4806 0.5967 0.7724
No log 3.4595 128 0.5472 0.5383 0.5472 0.7397
No log 3.5135 130 0.6402 0.5298 0.6402 0.8001
No log 3.5676 132 0.5955 0.5414 0.5955 0.7717
No log 3.6216 134 0.5479 0.5574 0.5479 0.7402
No log 3.6757 136 0.7060 0.4862 0.7060 0.8403
No log 3.7297 138 0.8519 0.4232 0.8519 0.9230
No log 3.7838 140 0.8188 0.4835 0.8188 0.9049
No log 3.8378 142 0.6571 0.4788 0.6571 0.8106
No log 3.8919 144 0.5447 0.5115 0.5447 0.7381
No log 3.9459 146 0.5310 0.5488 0.5310 0.7287
No log 4.0 148 0.5727 0.4737 0.5727 0.7567
No log 4.0541 150 0.5917 0.4836 0.5917 0.7692
No log 4.1081 152 0.5577 0.4562 0.5577 0.7468
No log 4.1622 154 0.5261 0.4972 0.5261 0.7253
No log 4.2162 156 0.5364 0.5208 0.5364 0.7324
No log 4.2703 158 0.5286 0.4762 0.5286 0.7271
No log 4.3243 160 0.5288 0.4762 0.5288 0.7272
No log 4.3784 162 0.5406 0.5356 0.5406 0.7352
No log 4.4324 164 0.5900 0.5763 0.5900 0.7681
No log 4.4865 166 0.6806 0.5003 0.6806 0.8250
No log 4.5405 168 0.6663 0.5460 0.6663 0.8163
No log 4.5946 170 0.5547 0.5483 0.5547 0.7448
No log 4.6486 172 0.5400 0.5452 0.5400 0.7349
No log 4.7027 174 0.5510 0.5178 0.5510 0.7423
No log 4.7568 176 0.5483 0.5178 0.5483 0.7404
No log 4.8108 178 0.5461 0.4147 0.5461 0.7390
No log 4.8649 180 0.6749 0.4307 0.6749 0.8215
No log 4.9189 182 0.6948 0.4307 0.6948 0.8335
No log 4.9730 184 0.5772 0.4430 0.5772 0.7598
No log 5.0270 186 0.5469 0.5098 0.5469 0.7395
No log 5.0811 188 0.5375 0.5682 0.5375 0.7332
No log 5.1351 190 0.5287 0.5476 0.5287 0.7271
No log 5.1892 192 0.5377 0.5753 0.5377 0.7333
No log 5.2432 194 0.7188 0.4085 0.7188 0.8478
No log 5.2973 196 0.9028 0.4503 0.9028 0.9502
No log 5.3514 198 0.8046 0.4044 0.8046 0.8970
No log 5.4054 200 0.6016 0.3746 0.6016 0.7756
No log 5.4595 202 0.5512 0.3536 0.5512 0.7424
No log 5.5135 204 0.5563 0.4083 0.5563 0.7458
No log 5.5676 206 0.5600 0.4463 0.5600 0.7483
No log 5.6216 208 0.6739 0.4230 0.6739 0.8209
No log 5.6757 210 0.7387 0.4230 0.7387 0.8595
No log 5.7297 212 0.6395 0.4582 0.6395 0.7997
No log 5.7838 214 0.5463 0.4801 0.5463 0.7392
No log 5.8378 216 0.5414 0.5248 0.5414 0.7358
No log 5.8919 218 0.5238 0.5463 0.5238 0.7237
No log 5.9459 220 0.5927 0.5093 0.5927 0.7699
No log 6.0 222 0.8006 0.4978 0.8006 0.8948
No log 6.0541 224 0.7909 0.4502 0.7909 0.8893
No log 6.1081 226 0.6777 0.4522 0.6777 0.8233
No log 6.1622 228 0.6025 0.4434 0.6025 0.7762
No log 6.2162 230 0.6335 0.5098 0.6335 0.7959
No log 6.2703 232 0.6958 0.5038 0.6958 0.8342
No log 6.3243 234 0.7092 0.5036 0.7092 0.8422
No log 6.3784 236 0.6273 0.5349 0.6273 0.7920
No log 6.4324 238 0.5959 0.4642 0.5959 0.7719
No log 6.4865 240 0.5721 0.4379 0.5721 0.7564
No log 6.5405 242 0.5476 0.5357 0.5476 0.7400
No log 6.5946 244 0.5471 0.5357 0.5471 0.7397
No log 6.6486 246 0.5457 0.5450 0.5457 0.7387
No log 6.7027 248 0.5480 0.5634 0.5480 0.7403
No log 6.7568 250 0.5537 0.6059 0.5537 0.7441
No log 6.8108 252 0.5565 0.6059 0.5565 0.7460
No log 6.8649 254 0.5593 0.5697 0.5593 0.7479
No log 6.9189 256 0.5933 0.5528 0.5933 0.7702
No log 6.9730 258 0.5379 0.5554 0.5379 0.7335
No log 7.0270 260 0.5152 0.5707 0.5152 0.7178
No log 7.0811 262 0.5844 0.4663 0.5844 0.7645
No log 7.1351 264 0.6462 0.4982 0.6462 0.8039
No log 7.1892 266 0.5628 0.5014 0.5628 0.7502
No log 7.2432 268 0.4974 0.6455 0.4974 0.7052
No log 7.2973 270 0.5019 0.6046 0.5019 0.7084
No log 7.3514 272 0.5521 0.5173 0.5521 0.7430
No log 7.4054 274 0.5978 0.4734 0.5978 0.7732
No log 7.4595 276 0.5744 0.4951 0.5744 0.7579
No log 7.5135 278 0.5293 0.5632 0.5293 0.7275
No log 7.5676 280 0.5198 0.5782 0.5198 0.7210
No log 7.6216 282 0.5247 0.5405 0.5247 0.7244
No log 7.6757 284 0.5485 0.5632 0.5485 0.7406
No log 7.7297 286 0.6146 0.5081 0.6146 0.7839
No log 7.7838 288 0.5825 0.5081 0.5825 0.7632
No log 7.8378 290 0.5643 0.5560 0.5643 0.7512
No log 7.8919 292 0.5449 0.5053 0.5449 0.7382
No log 7.9459 294 0.5619 0.5560 0.5619 0.7496
No log 8.0 296 0.6426 0.4808 0.6426 0.8016
No log 8.0541 298 0.6778 0.4502 0.6778 0.8233
No log 8.1081 300 0.5860 0.5206 0.5860 0.7655
No log 8.1622 302 0.5344 0.5071 0.5344 0.7310
No log 8.2162 304 0.5337 0.5195 0.5337 0.7305
No log 8.2703 306 0.5522 0.4384 0.5522 0.7431
No log 8.3243 308 0.5728 0.4101 0.5728 0.7568
No log 8.3784 310 0.5784 0.3558 0.5784 0.7605
No log 8.4324 312 0.5750 0.4101 0.5750 0.7583
No log 8.4865 314 0.5710 0.3481 0.5710 0.7556
No log 8.5405 316 0.5697 0.3728 0.5697 0.7548
No log 8.5946 318 0.5755 0.5056 0.5755 0.7586
No log 8.6486 320 0.5708 0.4569 0.5708 0.7555
No log 8.7027 322 0.5508 0.3995 0.5508 0.7422
No log 8.7568 324 0.5382 0.4484 0.5382 0.7336
No log 8.8108 326 0.5341 0.4575 0.5341 0.7309
No log 8.8649 328 0.5651 0.5647 0.5651 0.7517
No log 8.9189 330 0.5403 0.5768 0.5403 0.7351
No log 8.9730 332 0.5315 0.6254 0.5315 0.7290
No log 9.0270 334 0.5399 0.5289 0.5399 0.7348
No log 9.0811 336 0.5634 0.5289 0.5634 0.7506
No log 9.1351 338 0.5615 0.5289 0.5615 0.7493
No log 9.1892 340 0.5526 0.5687 0.5526 0.7434
No log 9.2432 342 0.5599 0.5738 0.5599 0.7482
No log 9.2973 344 0.5447 0.5951 0.5447 0.7381
No log 9.3514 346 0.5324 0.5738 0.5324 0.7297
No log 9.4054 348 0.5125 0.5460 0.5125 0.7159
No log 9.4595 350 0.5084 0.5089 0.5084 0.7130
No log 9.5135 352 0.5122 0.5609 0.5122 0.7157
No log 9.5676 354 0.5109 0.5177 0.5109 0.7148
No log 9.6216 356 0.5081 0.4923 0.5081 0.7128
No log 9.6757 358 0.5041 0.4757 0.5041 0.7100
No log 9.7297 360 0.5115 0.5816 0.5115 0.7152
No log 9.7838 362 0.5624 0.5237 0.5624 0.7499
No log 9.8378 364 0.5416 0.5752 0.5416 0.7359
No log 9.8919 366 0.5270 0.4596 0.5270 0.7260
No log 9.9459 368 0.5370 0.4705 0.5370 0.7328
No log 10.0 370 0.6074 0.4788 0.6074 0.7794
No log 10.0541 372 0.6272 0.4788 0.6272 0.7920
No log 10.1081 374 0.5604 0.4576 0.5604 0.7486
No log 10.1622 376 0.5397 0.4569 0.5397 0.7346
No log 10.2162 378 0.5458 0.4158 0.5458 0.7388
No log 10.2703 380 0.5971 0.4644 0.5971 0.7727
No log 10.3243 382 0.6081 0.4644 0.6081 0.7798
No log 10.3784 384 0.5654 0.4815 0.5654 0.7520
No log 10.4324 386 0.5499 0.4257 0.5499 0.7416
No log 10.4865 388 0.5545 0.4257 0.5545 0.7447
No log 10.5405 390 0.5760 0.5104 0.5760 0.7590
No log 10.5946 392 0.5636 0.4875 0.5636 0.7507
No log 10.6486 394 0.5426 0.4726 0.5426 0.7366
No log 10.7027 396 0.5448 0.5107 0.5448 0.7381
No log 10.7568 398 0.5453 0.5107 0.5453 0.7384
No log 10.8108 400 0.5500 0.3865 0.5500 0.7416
No log 10.8649 402 0.5696 0.3622 0.5696 0.7547
No log 10.9189 404 0.5973 0.4997 0.5973 0.7729
No log 10.9730 406 0.5472 0.3865 0.5472 0.7397
No log 11.0270 408 0.5145 0.4762 0.5145 0.7173
No log 11.0811 410 0.5134 0.4904 0.5134 0.7165
No log 11.1351 412 0.5472 0.4358 0.5472 0.7397
No log 11.1892 414 0.6722 0.4892 0.6722 0.8199
No log 11.2432 416 0.6945 0.4892 0.6945 0.8333
No log 11.2973 418 0.6212 0.4997 0.6212 0.7881
No log 11.3514 420 0.5307 0.4484 0.5307 0.7285
No log 11.4054 422 0.5258 0.4802 0.5258 0.7251
No log 11.4595 424 0.5266 0.4548 0.5266 0.7257
No log 11.5135 426 0.5399 0.4419 0.5399 0.7347
No log 11.5676 428 0.5864 0.4845 0.5864 0.7658
No log 11.6216 430 0.6411 0.4815 0.6411 0.8007
No log 11.6757 432 0.6266 0.4247 0.6266 0.7916
No log 11.7297 434 0.5752 0.4243 0.5752 0.7584
No log 11.7838 436 0.5526 0.4681 0.5526 0.7434
No log 11.8378 438 0.5493 0.4657 0.5493 0.7412
No log 11.8919 440 0.5644 0.5092 0.5644 0.7513
No log 11.9459 442 0.5592 0.5217 0.5592 0.7478
No log 12.0 444 0.5975 0.4919 0.5975 0.7730
No log 12.0541 446 0.6939 0.5309 0.6939 0.8330
No log 12.1081 448 0.7711 0.5199 0.7711 0.8781
No log 12.1622 450 0.7223 0.5387 0.7223 0.8499
No log 12.2162 452 0.6054 0.5310 0.6054 0.7780
No log 12.2703 454 0.5331 0.4801 0.5331 0.7302
No log 12.3243 456 0.5099 0.4801 0.5099 0.7141
No log 12.3784 458 0.5160 0.5937 0.5160 0.7183
No log 12.4324 460 0.5688 0.5950 0.5688 0.7542
No log 12.4865 462 0.5672 0.5950 0.5672 0.7532
No log 12.5405 464 0.5257 0.5141 0.5257 0.7250
No log 12.5946 466 0.5169 0.4657 0.5169 0.7189
No log 12.6486 468 0.5190 0.4448 0.5190 0.7204
No log 12.7027 470 0.5016 0.5436 0.5016 0.7082
No log 12.7568 472 0.4796 0.6065 0.4796 0.6925
No log 12.8108 474 0.5155 0.5770 0.5155 0.7180
No log 12.8649 476 0.5269 0.5872 0.5269 0.7259
No log 12.9189 478 0.4868 0.5697 0.4868 0.6977
No log 12.9730 480 0.4757 0.6656 0.4757 0.6897
No log 13.0270 482 0.5051 0.5697 0.5051 0.7107
No log 13.0811 484 0.4996 0.5563 0.4996 0.7068
No log 13.1351 486 0.4867 0.6241 0.4867 0.6976
No log 13.1892 488 0.6044 0.6078 0.6044 0.7774
No log 13.2432 490 0.7219 0.5195 0.7219 0.8496
No log 13.2973 492 0.6971 0.5195 0.6971 0.8350
No log 13.3514 494 0.5944 0.5646 0.5944 0.7709
No log 13.4054 496 0.5196 0.5560 0.5196 0.7209
No log 13.4595 498 0.4922 0.6241 0.4922 0.7016
0.3162 13.5135 500 0.4894 0.5549 0.4894 0.6996
0.3162 13.5676 502 0.5215 0.4795 0.5215 0.7221
0.3162 13.6216 504 0.5404 0.5335 0.5404 0.7351
0.3162 13.6757 506 0.5249 0.5075 0.5249 0.7245
0.3162 13.7297 508 0.5211 0.6252 0.5211 0.7219
0.3162 13.7838 510 0.6096 0.5388 0.6096 0.7808
0.3162 13.8378 512 0.6834 0.5443 0.6834 0.8267
0.3162 13.8919 514 0.6431 0.5008 0.6431 0.8020
0.3162 13.9459 516 0.5627 0.5237 0.5627 0.7501
0.3162 14.0 518 0.5474 0.5609 0.5474 0.7398
0.3162 14.0541 520 0.5469 0.5460 0.5469 0.7395
0.3162 14.1081 522 0.5513 0.5533 0.5513 0.7425
0.3162 14.1622 524 0.5989 0.4618 0.5989 0.7739
0.3162 14.2162 526 0.6875 0.5258 0.6875 0.8291
0.3162 14.2703 528 0.7143 0.4723 0.7143 0.8451

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k7_task7_organization

Finetuned
(4019)
this model