ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k2_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4575
  • Qwk: 0.5510
  • Mse: 0.4575
  • Rmse: 0.6764

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 2.6186 -0.0230 2.6186 1.6182
No log 0.4 4 1.4315 0.0763 1.4315 1.1964
No log 0.6 6 0.7538 0.1321 0.7538 0.8682
No log 0.8 8 0.7517 0.1328 0.7517 0.8670
No log 1.0 10 0.6860 0.2621 0.6860 0.8283
No log 1.2 12 0.6510 0.4235 0.6510 0.8068
No log 1.4 14 0.7018 0.4629 0.7018 0.8377
No log 1.6 16 0.6685 0.4788 0.6685 0.8176
No log 1.8 18 0.5880 0.4504 0.5880 0.7668
No log 2.0 20 0.5781 0.4229 0.5781 0.7603
No log 2.2 22 0.5424 0.4591 0.5424 0.7365
No log 2.4 24 0.5957 0.4556 0.5957 0.7718
No log 2.6 26 0.7313 0.5093 0.7313 0.8552
No log 2.8 28 0.6490 0.5423 0.6490 0.8056
No log 3.0 30 0.4591 0.6269 0.4591 0.6776
No log 3.2 32 0.5140 0.4724 0.5140 0.7169
No log 3.4 34 0.4563 0.6140 0.4563 0.6755
No log 3.6 36 0.7653 0.4815 0.7653 0.8748
No log 3.8 38 0.8689 0.3359 0.8689 0.9321
No log 4.0 40 0.7196 0.5706 0.7196 0.8483
No log 4.2 42 0.7144 0.2986 0.7144 0.8452
No log 4.4 44 1.1323 0.2037 1.1323 1.0641
No log 4.6 46 0.9095 0.3728 0.9095 0.9537
No log 4.8 48 0.4619 0.5214 0.4619 0.6796
No log 5.0 50 0.4460 0.5633 0.4460 0.6678
No log 5.2 52 0.6082 0.5778 0.6082 0.7799
No log 5.4 54 0.6113 0.5560 0.6113 0.7819
No log 5.6 56 0.4764 0.4958 0.4764 0.6902
No log 5.8 58 0.4724 0.5237 0.4724 0.6873
No log 6.0 60 0.6527 0.5237 0.6527 0.8079
No log 6.2 62 0.5431 0.6110 0.5431 0.7369
No log 6.4 64 0.4447 0.6337 0.4447 0.6669
No log 6.6 66 0.5622 0.5340 0.5622 0.7498
No log 6.8 68 0.6671 0.5354 0.6671 0.8168
No log 7.0 70 0.7020 0.5429 0.7020 0.8378
No log 7.2 72 0.5257 0.5528 0.5257 0.7251
No log 7.4 74 0.4369 0.5875 0.4369 0.6610
No log 7.6 76 0.4841 0.6284 0.4841 0.6958
No log 7.8 78 0.5183 0.5946 0.5183 0.7200
No log 8.0 80 0.4934 0.4278 0.4934 0.7024
No log 8.2 82 0.4986 0.4742 0.4986 0.7061
No log 8.4 84 0.4732 0.5021 0.4732 0.6879
No log 8.6 86 0.4486 0.6265 0.4486 0.6698
No log 8.8 88 0.4571 0.5494 0.4571 0.6761
No log 9.0 90 0.4769 0.6620 0.4769 0.6906
No log 9.2 92 0.4654 0.7051 0.4654 0.6822
No log 9.4 94 0.4414 0.6395 0.4414 0.6644
No log 9.6 96 0.4662 0.6880 0.4662 0.6828
No log 9.8 98 0.4398 0.6371 0.4398 0.6632
No log 10.0 100 0.4376 0.6479 0.4376 0.6615
No log 10.2 102 0.4458 0.6118 0.4458 0.6677
No log 10.4 104 0.4583 0.5649 0.4583 0.6770
No log 10.6 106 0.4698 0.6507 0.4698 0.6854
No log 10.8 108 0.4869 0.6507 0.4869 0.6978
No log 11.0 110 0.4907 0.6670 0.4907 0.7005
No log 11.2 112 0.4772 0.6227 0.4772 0.6908
No log 11.4 114 0.4563 0.6142 0.4563 0.6755
No log 11.6 116 0.4739 0.5611 0.4739 0.6884
No log 11.8 118 0.4800 0.5826 0.4800 0.6928
No log 12.0 120 0.4826 0.5632 0.4826 0.6947
No log 12.2 122 0.5566 0.5223 0.5566 0.7460
No log 12.4 124 0.5511 0.4721 0.5511 0.7423
No log 12.6 126 0.4698 0.6455 0.4698 0.6854
No log 12.8 128 0.5039 0.6431 0.5039 0.7098
No log 13.0 130 0.4903 0.6434 0.4903 0.7002
No log 13.2 132 0.4889 0.5872 0.4889 0.6992
No log 13.4 134 0.5129 0.6028 0.5129 0.7162
No log 13.6 136 0.4757 0.6441 0.4757 0.6897
No log 13.8 138 0.4713 0.5722 0.4713 0.6865
No log 14.0 140 0.4724 0.5463 0.4724 0.6873
No log 14.2 142 0.4579 0.6265 0.4579 0.6767
No log 14.4 144 0.5221 0.5576 0.5221 0.7226
No log 14.6 146 0.5535 0.5373 0.5535 0.7440
No log 14.8 148 0.4682 0.5703 0.4682 0.6843
No log 15.0 150 0.4794 0.4960 0.4794 0.6924
No log 15.2 152 0.4637 0.5234 0.4637 0.6810
No log 15.4 154 0.4632 0.6228 0.4632 0.6806
No log 15.6 156 0.4833 0.6038 0.4833 0.6952
No log 15.8 158 0.4953 0.6025 0.4953 0.7038
No log 16.0 160 0.4583 0.6346 0.4583 0.6770
No log 16.2 162 0.4553 0.5993 0.4553 0.6747
No log 16.4 164 0.4522 0.6552 0.4522 0.6724
No log 16.6 166 0.4551 0.6541 0.4551 0.6746
No log 16.8 168 0.4502 0.6723 0.4502 0.6710
No log 17.0 170 0.4451 0.6455 0.4451 0.6672
No log 17.2 172 0.4420 0.6455 0.4420 0.6649
No log 17.4 174 0.4606 0.6341 0.4606 0.6787
No log 17.6 176 0.4672 0.6341 0.4672 0.6835
No log 17.8 178 0.4460 0.6275 0.4460 0.6678
No log 18.0 180 0.4502 0.5406 0.4502 0.6709
No log 18.2 182 0.4827 0.6154 0.4827 0.6947
No log 18.4 184 0.4554 0.5663 0.4554 0.6748
No log 18.6 186 0.4387 0.5902 0.4387 0.6624
No log 18.8 188 0.4533 0.6214 0.4533 0.6733
No log 19.0 190 0.4828 0.5219 0.4828 0.6948
No log 19.2 192 0.4729 0.5681 0.4729 0.6877
No log 19.4 194 0.4460 0.5309 0.4460 0.6678
No log 19.6 196 0.4419 0.5649 0.4419 0.6648
No log 19.8 198 0.5122 0.5407 0.5122 0.7157
No log 20.0 200 0.4668 0.5619 0.4668 0.6833
No log 20.2 202 0.4107 0.6073 0.4107 0.6408
No log 20.4 204 0.4162 0.6620 0.4162 0.6451
No log 20.6 206 0.4765 0.5880 0.4765 0.6903
No log 20.8 208 0.4545 0.6201 0.4545 0.6742
No log 21.0 210 0.4204 0.7012 0.4204 0.6484
No log 21.2 212 0.4151 0.6455 0.4151 0.6443
No log 21.4 214 0.4156 0.6455 0.4156 0.6447
No log 21.6 216 0.4135 0.6267 0.4135 0.6430
No log 21.8 218 0.4087 0.6277 0.4087 0.6393
No log 22.0 220 0.4056 0.6730 0.4056 0.6369
No log 22.2 222 0.4229 0.6993 0.4229 0.6503
No log 22.4 224 0.4446 0.6608 0.4446 0.6668
No log 22.6 226 0.4114 0.7033 0.4114 0.6414
No log 22.8 228 0.4364 0.4958 0.4364 0.6606
No log 23.0 230 0.4592 0.5301 0.4592 0.6776
No log 23.2 232 0.4391 0.5208 0.4391 0.6626
No log 23.4 234 0.4215 0.6277 0.4215 0.6492
No log 23.6 236 0.4311 0.6818 0.4311 0.6566
No log 23.8 238 0.4387 0.6993 0.4387 0.6624
No log 24.0 240 0.4293 0.6277 0.4293 0.6552
No log 24.2 242 0.4328 0.5421 0.4328 0.6578
No log 24.4 244 0.4294 0.5373 0.4294 0.6553
No log 24.6 246 0.4322 0.6351 0.4322 0.6574
No log 24.8 248 0.4323 0.6351 0.4323 0.6575
No log 25.0 250 0.4306 0.6156 0.4306 0.6562
No log 25.2 252 0.4292 0.5956 0.4292 0.6551
No log 25.4 254 0.4285 0.6242 0.4285 0.6546
No log 25.6 256 0.4464 0.5868 0.4464 0.6681
No log 25.8 258 0.4532 0.5868 0.4532 0.6732
No log 26.0 260 0.4458 0.5648 0.4458 0.6676
No log 26.2 262 0.4422 0.6096 0.4422 0.6650
No log 26.4 264 0.4363 0.6639 0.4363 0.6605
No log 26.6 266 0.4393 0.5669 0.4393 0.6628
No log 26.8 268 0.4494 0.5452 0.4494 0.6704
No log 27.0 270 0.4311 0.5414 0.4311 0.6566
No log 27.2 272 0.4456 0.5819 0.4456 0.6675
No log 27.4 274 0.4821 0.5219 0.4821 0.6943
No log 27.6 276 0.5087 0.5345 0.5087 0.7132
No log 27.8 278 0.4810 0.6368 0.4810 0.6936
No log 28.0 280 0.4279 0.6807 0.4279 0.6542
No log 28.2 282 0.4081 0.6073 0.4081 0.6389
No log 28.4 284 0.4158 0.5463 0.4158 0.6449
No log 28.6 286 0.4182 0.5414 0.4182 0.6467
No log 28.8 288 0.4144 0.5414 0.4144 0.6438
No log 29.0 290 0.4195 0.5406 0.4195 0.6477
No log 29.2 292 0.4175 0.6551 0.4175 0.6462
No log 29.4 294 0.4267 0.7118 0.4267 0.6532
No log 29.6 296 0.4383 0.7097 0.4383 0.6620
No log 29.8 298 0.4254 0.7118 0.4254 0.6523
No log 30.0 300 0.4147 0.6554 0.4147 0.6440
No log 30.2 302 0.4136 0.6739 0.4136 0.6431
No log 30.4 304 0.4141 0.6739 0.4141 0.6435
No log 30.6 306 0.4145 0.6739 0.4145 0.6438
No log 30.8 308 0.4167 0.6739 0.4167 0.6455
No log 31.0 310 0.4214 0.6364 0.4214 0.6491
No log 31.2 312 0.4299 0.5373 0.4299 0.6557
No log 31.4 314 0.4356 0.5321 0.4356 0.6600
No log 31.6 316 0.4371 0.5373 0.4371 0.6611
No log 31.8 318 0.4474 0.6890 0.4474 0.6689
No log 32.0 320 0.4571 0.6495 0.4571 0.6761
No log 32.2 322 0.4470 0.7044 0.4470 0.6686
No log 32.4 324 0.4302 0.6975 0.4302 0.6559
No log 32.6 326 0.4218 0.6994 0.4218 0.6495
No log 32.8 328 0.4483 0.6066 0.4483 0.6696
No log 33.0 330 0.4573 0.5947 0.4573 0.6763
No log 33.2 332 0.4634 0.5921 0.4634 0.6808
No log 33.4 334 0.4257 0.5390 0.4257 0.6525
No log 33.6 336 0.4140 0.6727 0.4140 0.6434
No log 33.8 338 0.4262 0.6807 0.4262 0.6528
No log 34.0 340 0.4310 0.6807 0.4310 0.6565
No log 34.2 342 0.4345 0.6616 0.4345 0.6592
No log 34.4 344 0.4376 0.5397 0.4376 0.6615
No log 34.6 346 0.4486 0.5030 0.4486 0.6697
No log 34.8 348 0.4523 0.5098 0.4523 0.6725
No log 35.0 350 0.4578 0.5463 0.4578 0.6766
No log 35.2 352 0.4513 0.5125 0.4513 0.6718
No log 35.4 354 0.4483 0.5816 0.4483 0.6695
No log 35.6 356 0.4533 0.6147 0.4533 0.6733
No log 35.8 358 0.4411 0.6623 0.4411 0.6641
No log 36.0 360 0.4294 0.5812 0.4294 0.6553
No log 36.2 362 0.4275 0.5580 0.4275 0.6538
No log 36.4 364 0.4401 0.5248 0.4401 0.6634
No log 36.6 366 0.4440 0.5422 0.4440 0.6663
No log 36.8 368 0.4240 0.5479 0.4240 0.6511
No log 37.0 370 0.4134 0.6458 0.4134 0.6430
No log 37.2 372 0.4101 0.6904 0.4101 0.6404
No log 37.4 374 0.4107 0.6830 0.4107 0.6409
No log 37.6 376 0.4112 0.6830 0.4112 0.6413
No log 37.8 378 0.4106 0.6830 0.4106 0.6407
No log 38.0 380 0.4391 0.5266 0.4391 0.6626
No log 38.2 382 0.4874 0.5015 0.4874 0.6981
No log 38.4 384 0.4944 0.5015 0.4944 0.7031
No log 38.6 386 0.4604 0.5226 0.4604 0.6785
No log 38.8 388 0.4251 0.5538 0.4251 0.6520
No log 39.0 390 0.4386 0.6687 0.4386 0.6623
No log 39.2 392 0.4938 0.6577 0.4938 0.7027
No log 39.4 394 0.5252 0.6214 0.5252 0.7247
No log 39.6 396 0.4919 0.6577 0.4919 0.7013
No log 39.8 398 0.4387 0.6786 0.4387 0.6623
No log 40.0 400 0.4071 0.6060 0.4071 0.6380
No log 40.2 402 0.4145 0.5596 0.4145 0.6438
No log 40.4 404 0.4169 0.5596 0.4169 0.6457
No log 40.6 406 0.4143 0.5596 0.4143 0.6436
No log 40.8 408 0.4130 0.6279 0.4130 0.6426
No log 41.0 410 0.4144 0.6156 0.4144 0.6437
No log 41.2 412 0.4263 0.6326 0.4263 0.6530
No log 41.4 414 0.4371 0.6187 0.4371 0.6612
No log 41.6 416 0.4409 0.5980 0.4409 0.6640
No log 41.8 418 0.4362 0.5980 0.4362 0.6604
No log 42.0 420 0.4319 0.4972 0.4319 0.6572
No log 42.2 422 0.4435 0.5463 0.4435 0.6660
No log 42.4 424 0.4632 0.5406 0.4632 0.6806
No log 42.6 426 0.4529 0.5611 0.4529 0.6730
No log 42.8 428 0.4300 0.5373 0.4300 0.6557
No log 43.0 430 0.4196 0.5373 0.4196 0.6478
No log 43.2 432 0.4155 0.6156 0.4155 0.6446
No log 43.4 434 0.4201 0.5446 0.4201 0.6481
No log 43.6 436 0.4258 0.5397 0.4258 0.6525
No log 43.8 438 0.4303 0.5152 0.4303 0.6560
No log 44.0 440 0.4322 0.5213 0.4322 0.6574
No log 44.2 442 0.4375 0.5941 0.4375 0.6614
No log 44.4 444 0.4418 0.6284 0.4418 0.6647
No log 44.6 446 0.4418 0.6295 0.4418 0.6647
No log 44.8 448 0.4503 0.5519 0.4503 0.6710
No log 45.0 450 0.4619 0.5628 0.4619 0.6796
No log 45.2 452 0.4764 0.5317 0.4764 0.6902
No log 45.4 454 0.4592 0.5117 0.4592 0.6776
No log 45.6 456 0.4410 0.5028 0.4410 0.6641
No log 45.8 458 0.4333 0.5812 0.4333 0.6583
No log 46.0 460 0.4383 0.6975 0.4383 0.6620
No log 46.2 462 0.4468 0.6414 0.4468 0.6684
No log 46.4 464 0.4411 0.6426 0.4411 0.6641
No log 46.6 466 0.4369 0.5815 0.4369 0.6610
No log 46.8 468 0.4396 0.5030 0.4396 0.6630
No log 47.0 470 0.4356 0.5030 0.4356 0.6600
No log 47.2 472 0.4311 0.5538 0.4311 0.6566
No log 47.4 474 0.4304 0.5446 0.4304 0.6561
No log 47.6 476 0.4355 0.6643 0.4355 0.6599
No log 47.8 478 0.4409 0.6904 0.4409 0.6640
No log 48.0 480 0.4497 0.5970 0.4497 0.6706
No log 48.2 482 0.4547 0.6040 0.4547 0.6743
No log 48.4 484 0.4560 0.5866 0.4560 0.6753
No log 48.6 486 0.4523 0.5492 0.4523 0.6725
No log 48.8 488 0.4512 0.5538 0.4512 0.6717
No log 49.0 490 0.4530 0.5246 0.4530 0.6730
No log 49.2 492 0.4519 0.4990 0.4519 0.6722
No log 49.4 494 0.4484 0.4990 0.4484 0.6696
No log 49.6 496 0.4440 0.5304 0.4440 0.6663
No log 49.8 498 0.4441 0.5446 0.4441 0.6664
0.2245 50.0 500 0.4484 0.6198 0.4484 0.6696
0.2245 50.2 502 0.4546 0.6706 0.4546 0.6743
0.2245 50.4 504 0.4506 0.6630 0.4506 0.6713
0.2245 50.6 506 0.4459 0.6198 0.4459 0.6678
0.2245 50.8 508 0.4423 0.5619 0.4423 0.6650
0.2245 51.0 510 0.4509 0.5463 0.4509 0.6715
0.2245 51.2 512 0.4572 0.5555 0.4572 0.6761
0.2245 51.4 514 0.4575 0.5510 0.4575 0.6764

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k2_task7_organization

Finetuned
(4019)
this model