ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k9_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5516
  • Qwk: 0.4802
  • Mse: 0.5516
  • Rmse: 0.7427

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0426 2 2.4856 -0.0788 2.4856 1.5766
No log 0.0851 4 1.3022 0.0998 1.3022 1.1411
No log 0.1277 6 0.7469 0.0481 0.7469 0.8642
No log 0.1702 8 0.8475 0.1822 0.8475 0.9206
No log 0.2128 10 1.0275 0.2589 1.0275 1.0137
No log 0.2553 12 0.8001 0.1800 0.8001 0.8945
No log 0.2979 14 0.8666 0.1304 0.8666 0.9309
No log 0.3404 16 0.7610 0.1687 0.7610 0.8724
No log 0.3830 18 0.7281 0.1228 0.7281 0.8533
No log 0.4255 20 0.7851 0.2395 0.7851 0.8861
No log 0.4681 22 1.0522 0.1528 1.0522 1.0258
No log 0.5106 24 1.4446 0.0372 1.4446 1.2019
No log 0.5532 26 1.3476 0.0785 1.3476 1.1609
No log 0.5957 28 1.3835 0.0789 1.3835 1.1762
No log 0.6383 30 1.4728 0.0182 1.4728 1.2136
No log 0.6809 32 1.2393 0.1686 1.2393 1.1133
No log 0.7234 34 0.9263 0.2508 0.9263 0.9624
No log 0.7660 36 0.7733 0.2424 0.7733 0.8794
No log 0.8085 38 0.7799 0.2156 0.7799 0.8831
No log 0.8511 40 0.8235 0.2879 0.8235 0.9075
No log 0.8936 42 0.9435 0.3598 0.9435 0.9713
No log 0.9362 44 0.9604 0.3875 0.9604 0.9800
No log 0.9787 46 0.8773 0.3090 0.8773 0.9366
No log 1.0213 48 0.8737 0.3051 0.8737 0.9347
No log 1.0638 50 0.7693 0.2841 0.7693 0.8771
No log 1.1064 52 0.6956 0.2036 0.6956 0.8340
No log 1.1489 54 0.6519 0.2418 0.6519 0.8074
No log 1.1915 56 0.6462 0.2389 0.6462 0.8039
No log 1.2340 58 0.6743 0.2459 0.6743 0.8212
No log 1.2766 60 0.6873 0.2430 0.6873 0.8290
No log 1.3191 62 0.6598 0.2424 0.6598 0.8123
No log 1.3617 64 0.6573 0.2940 0.6573 0.8107
No log 1.4043 66 0.6782 0.2833 0.6782 0.8235
No log 1.4468 68 0.6246 0.3834 0.6246 0.7903
No log 1.4894 70 0.7150 0.3614 0.7150 0.8456
No log 1.5319 72 1.0238 0.1959 1.0238 1.0118
No log 1.5745 74 1.3760 0.1199 1.3760 1.1730
No log 1.6170 76 1.3410 0.0983 1.3410 1.1580
No log 1.6596 78 0.9821 0.1760 0.9821 0.9910
No log 1.7021 80 0.5755 0.4681 0.5755 0.7587
No log 1.7447 82 0.7785 0.3347 0.7785 0.8823
No log 1.7872 84 0.8642 0.3516 0.8642 0.9296
No log 1.8298 86 0.6913 0.3894 0.6913 0.8314
No log 1.8723 88 0.5583 0.3006 0.5583 0.7472
No log 1.9149 90 0.7202 0.2804 0.7202 0.8487
No log 1.9574 92 1.0287 0.2701 1.0287 1.0142
No log 2.0 94 1.2009 0.1882 1.2009 1.0959
No log 2.0426 96 1.2682 0.1439 1.2682 1.1261
No log 2.0851 98 1.1590 0.1903 1.1590 1.0766
No log 2.1277 100 0.7520 0.4186 0.7520 0.8672
No log 2.1702 102 0.5522 0.4788 0.5522 0.7431
No log 2.2128 104 0.5790 0.4378 0.5790 0.7609
No log 2.2553 106 0.6481 0.4291 0.6481 0.8051
No log 2.2979 108 0.6986 0.4267 0.6986 0.8358
No log 2.3404 110 0.6189 0.3336 0.6189 0.7867
No log 2.3830 112 0.6470 0.5016 0.6470 0.8043
No log 2.4255 114 0.6700 0.4451 0.6700 0.8185
No log 2.4681 116 0.5870 0.5016 0.5870 0.7662
No log 2.5106 118 0.5301 0.6122 0.5301 0.7281
No log 2.5532 120 0.5310 0.5631 0.5310 0.7287
No log 2.5957 122 0.5236 0.6230 0.5236 0.7236
No log 2.6383 124 0.5736 0.4949 0.5736 0.7574
No log 2.6809 126 0.5633 0.5368 0.5633 0.7505
No log 2.7234 128 0.5407 0.5918 0.5407 0.7353
No log 2.7660 130 0.5529 0.5554 0.5529 0.7435
No log 2.8085 132 0.6727 0.4332 0.6727 0.8202
No log 2.8511 134 0.8381 0.4078 0.8381 0.9155
No log 2.8936 136 0.9786 0.3275 0.9786 0.9892
No log 2.9362 138 0.9321 0.3553 0.9321 0.9655
No log 2.9787 140 0.7391 0.4683 0.7391 0.8597
No log 3.0213 142 0.6205 0.4850 0.6205 0.7877
No log 3.0638 144 0.6433 0.5170 0.6433 0.8021
No log 3.1064 146 0.6426 0.4875 0.6426 0.8016
No log 3.1489 148 0.5656 0.5533 0.5656 0.7520
No log 3.1915 150 0.6467 0.5030 0.6467 0.8042
No log 3.2340 152 0.9081 0.2616 0.9081 0.9530
No log 3.2766 154 0.9644 0.2162 0.9644 0.9821
No log 3.3191 156 0.7986 0.3690 0.7986 0.8936
No log 3.3617 158 0.5856 0.5213 0.5856 0.7652
No log 3.4043 160 0.6001 0.4788 0.6001 0.7747
No log 3.4468 162 0.5882 0.5723 0.5882 0.7670
No log 3.4894 164 0.6070 0.5120 0.6070 0.7791
No log 3.5319 166 0.6860 0.4168 0.6860 0.8282
No log 3.5745 168 0.6411 0.4670 0.6411 0.8007
No log 3.6170 170 0.5655 0.56 0.5655 0.7520
No log 3.6596 172 0.5396 0.6108 0.5396 0.7346
No log 3.7021 174 0.5312 0.5860 0.5312 0.7288
No log 3.7447 176 0.5264 0.6129 0.5264 0.7255
No log 3.7872 178 0.5206 0.6129 0.5206 0.7215
No log 3.8298 180 0.5067 0.6254 0.5067 0.7119
No log 3.8723 182 0.4989 0.6184 0.4989 0.7063
No log 3.9149 184 0.5052 0.6566 0.5052 0.7108
No log 3.9574 186 0.4972 0.6142 0.4972 0.7051
No log 4.0 188 0.5028 0.6452 0.5028 0.7091
No log 4.0426 190 0.5174 0.6269 0.5174 0.7193
No log 4.0851 192 0.5196 0.6024 0.5196 0.7208
No log 4.1277 194 0.6000 0.3746 0.6000 0.7746
No log 4.1702 196 0.6830 0.3918 0.6830 0.8265
No log 4.2128 198 0.6287 0.3746 0.6287 0.7929
No log 4.2553 200 0.5646 0.5428 0.5646 0.7514
No log 4.2979 202 0.5680 0.6267 0.5680 0.7537
No log 4.3404 204 0.6293 0.5028 0.6293 0.7933
No log 4.3830 206 0.7198 0.4142 0.7198 0.8484
No log 4.4255 208 0.8481 0.3353 0.8481 0.9209
No log 4.4681 210 0.8097 0.3185 0.8097 0.8999
No log 4.5106 212 0.6813 0.5249 0.6813 0.8254
No log 4.5532 214 0.5699 0.5750 0.5699 0.7549
No log 4.5957 216 0.5762 0.3737 0.5762 0.7591
No log 4.6383 218 0.5721 0.3737 0.5721 0.7564
No log 4.6809 220 0.5446 0.4206 0.5446 0.7380
No log 4.7234 222 0.5637 0.4970 0.5637 0.7508
No log 4.7660 224 0.6420 0.4776 0.6420 0.8012
No log 4.8085 226 0.6440 0.4575 0.6440 0.8025
No log 4.8511 228 0.5724 0.5406 0.5724 0.7566
No log 4.8936 230 0.5417 0.5232 0.5417 0.7360
No log 4.9362 232 0.5905 0.4197 0.5905 0.7684
No log 4.9787 234 0.5661 0.4350 0.5661 0.7524
No log 5.0213 236 0.5344 0.5505 0.5344 0.7310
No log 5.0638 238 0.5586 0.5731 0.5586 0.7474
No log 5.1064 240 0.6242 0.4051 0.6242 0.7901
No log 5.1489 242 0.6159 0.4613 0.6159 0.7848
No log 5.1915 244 0.6615 0.4051 0.6615 0.8133
No log 5.2340 246 0.7823 0.3355 0.7823 0.8845
No log 5.2766 248 0.8668 0.3309 0.8668 0.9310
No log 5.3191 250 0.7555 0.3538 0.7555 0.8692
No log 5.3617 252 0.6119 0.3348 0.6119 0.7822
No log 5.4043 254 0.5682 0.3323 0.5682 0.7538
No log 5.4468 256 0.5646 0.3166 0.5646 0.7514
No log 5.4894 258 0.5822 0.2996 0.5822 0.7630
No log 5.5319 260 0.6129 0.3690 0.6129 0.7829
No log 5.5745 262 0.6057 0.3690 0.6057 0.7783
No log 5.6170 264 0.5768 0.2955 0.5768 0.7595
No log 5.6596 266 0.5597 0.3111 0.5597 0.7481
No log 5.7021 268 0.5739 0.3754 0.5739 0.7576
No log 5.7447 270 0.6813 0.4341 0.6813 0.8254
No log 5.7872 272 0.7873 0.3766 0.7873 0.8873
No log 5.8298 274 0.7303 0.4208 0.7303 0.8546
No log 5.8723 276 0.5711 0.4480 0.5711 0.7557
No log 5.9149 278 0.5676 0.4867 0.5676 0.7534
No log 5.9574 280 0.6841 0.3069 0.6841 0.8271
No log 6.0 282 0.7512 0.3294 0.7512 0.8667
No log 6.0426 284 0.6967 0.3234 0.6967 0.8347
No log 6.0851 286 0.5970 0.3662 0.5970 0.7726
No log 6.1277 288 0.5666 0.5213 0.5666 0.7527
No log 6.1702 290 0.5805 0.5933 0.5805 0.7619
No log 6.2128 292 0.5674 0.5250 0.5674 0.7533
No log 6.2553 294 0.5548 0.5421 0.5548 0.7448
No log 6.2979 296 0.5536 0.5476 0.5536 0.7441
No log 6.3404 298 0.5508 0.5768 0.5508 0.7421
No log 6.3830 300 0.5526 0.5840 0.5526 0.7434
No log 6.4255 302 0.5520 0.5840 0.5520 0.7430
No log 6.4681 304 0.5552 0.5840 0.5552 0.7451
No log 6.5106 306 0.5658 0.6223 0.5658 0.7522
No log 6.5532 308 0.6006 0.4321 0.6006 0.7750
No log 6.5957 310 0.6012 0.4732 0.6012 0.7753
No log 6.6383 312 0.5755 0.4958 0.5755 0.7586
No log 6.6809 314 0.5708 0.4389 0.5708 0.7555
No log 6.7234 316 0.5872 0.3804 0.5872 0.7663
No log 6.7660 318 0.5944 0.4293 0.5944 0.7710
No log 6.8085 320 0.5681 0.5390 0.5681 0.7537
No log 6.8511 322 0.5567 0.4678 0.5567 0.7461
No log 6.8936 324 0.5613 0.4422 0.5613 0.7492
No log 6.9362 326 0.5554 0.4713 0.5554 0.7453
No log 6.9787 328 0.5495 0.5254 0.5495 0.7413
No log 7.0213 330 0.5415 0.6083 0.5415 0.7358
No log 7.0638 332 0.5334 0.6083 0.5334 0.7303
No log 7.1064 334 0.5369 0.5798 0.5369 0.7327
No log 7.1489 336 0.6094 0.4210 0.6094 0.7807
No log 7.1915 338 0.6980 0.4930 0.6980 0.8355
No log 7.2340 340 0.7469 0.4946 0.7469 0.8642
No log 7.2766 342 0.6774 0.5061 0.6774 0.8230
No log 7.3191 344 0.6026 0.4544 0.6026 0.7763
No log 7.3617 346 0.5664 0.4302 0.5664 0.7526
No log 7.4043 348 0.6126 0.4706 0.6126 0.7827
No log 7.4468 350 0.7637 0.4472 0.7637 0.8739
No log 7.4894 352 0.8335 0.4683 0.8335 0.9130
No log 7.5319 354 0.7688 0.4426 0.7688 0.8768
No log 7.5745 356 0.6300 0.5439 0.6300 0.7937
No log 7.6170 358 0.6237 0.5439 0.6237 0.7898
No log 7.6596 360 0.6921 0.4686 0.6921 0.8319
No log 7.7021 362 0.7861 0.4367 0.7861 0.8866
No log 7.7447 364 0.7508 0.4116 0.7508 0.8665
No log 7.7872 366 0.6314 0.4653 0.6314 0.7946
No log 7.8298 368 0.5411 0.5436 0.5411 0.7356
No log 7.8723 370 0.5409 0.5373 0.5409 0.7355
No log 7.9149 372 0.5693 0.4813 0.5693 0.7545
No log 7.9574 374 0.6280 0.4765 0.6280 0.7924
No log 8.0 376 0.5933 0.5014 0.5933 0.7703
No log 8.0426 378 0.5232 0.4972 0.5232 0.7233
No log 8.0851 380 0.5261 0.4962 0.5261 0.7253
No log 8.1277 382 0.5236 0.4962 0.5236 0.7236
No log 8.1702 384 0.5486 0.4569 0.5486 0.7407
No log 8.2128 386 0.5940 0.4109 0.5940 0.7707
No log 8.2553 388 0.6034 0.5168 0.6034 0.7768
No log 8.2979 390 0.6380 0.5106 0.6380 0.7988
No log 8.3404 392 0.6721 0.4740 0.6721 0.8198
No log 8.3830 394 0.7040 0.5088 0.7040 0.8391
No log 8.4255 396 0.6332 0.4474 0.6332 0.7957
No log 8.4681 398 0.5765 0.5041 0.5765 0.7593
No log 8.5106 400 0.5410 0.4934 0.5410 0.7355
No log 8.5532 402 0.5316 0.4747 0.5316 0.7291
No log 8.5957 404 0.5242 0.4444 0.5242 0.7240
No log 8.6383 406 0.5347 0.4150 0.5347 0.7312
No log 8.6809 408 0.5870 0.4841 0.5870 0.7662
No log 8.7234 410 0.5820 0.5449 0.5820 0.7629
No log 8.7660 412 0.5301 0.4958 0.5301 0.7281
No log 8.8085 414 0.4916 0.5379 0.4916 0.7011
No log 8.8511 416 0.4908 0.5379 0.4908 0.7006
No log 8.8936 418 0.5159 0.5702 0.5159 0.7182
No log 8.9362 420 0.6154 0.4909 0.6154 0.7845
No log 8.9787 422 0.6927 0.4142 0.6927 0.8323
No log 9.0213 424 0.7047 0.4364 0.7047 0.8395
No log 9.0638 426 0.6176 0.4909 0.6176 0.7859
No log 9.1064 428 0.5198 0.5627 0.5198 0.7210
No log 9.1489 430 0.4784 0.5681 0.4784 0.6917
No log 9.1915 432 0.4853 0.5379 0.4853 0.6966
No log 9.2340 434 0.4862 0.6053 0.4862 0.6973
No log 9.2766 436 0.5521 0.5422 0.5521 0.7430
No log 9.3191 438 0.7081 0.4096 0.7081 0.8415
No log 9.3617 440 0.7501 0.4096 0.7501 0.8661
No log 9.4043 442 0.6957 0.4297 0.6957 0.8341
No log 9.4468 444 0.6149 0.4747 0.6149 0.7842
No log 9.4894 446 0.5597 0.5495 0.5597 0.7481
No log 9.5319 448 0.5283 0.5765 0.5283 0.7269
No log 9.5745 450 0.5263 0.5555 0.5263 0.7255
No log 9.6170 452 0.5407 0.5266 0.5407 0.7353
No log 9.6596 454 0.5990 0.5217 0.5990 0.7739
No log 9.7021 456 0.6075 0.5217 0.6075 0.7794
No log 9.7447 458 0.6427 0.4690 0.6427 0.8017
No log 9.7872 460 0.7055 0.5003 0.7055 0.8399
No log 9.8298 462 0.6455 0.4690 0.6455 0.8034
No log 9.8723 464 0.5578 0.4876 0.5578 0.7468
No log 9.9149 466 0.5375 0.4253 0.5375 0.7332
No log 9.9574 468 0.5408 0.3702 0.5408 0.7354
No log 10.0 470 0.5530 0.3702 0.5530 0.7436
No log 10.0426 472 0.6056 0.4174 0.6056 0.7782
No log 10.0851 474 0.6716 0.3661 0.6716 0.8195
No log 10.1277 476 0.7446 0.4258 0.7446 0.8629
No log 10.1702 478 0.7222 0.4258 0.7222 0.8498
No log 10.2128 480 0.6136 0.4711 0.6136 0.7833
No log 10.2553 482 0.5349 0.5379 0.5349 0.7314
No log 10.2979 484 0.5241 0.4722 0.5241 0.7240
No log 10.3404 486 0.5408 0.5781 0.5408 0.7354
No log 10.3830 488 0.6350 0.5030 0.6350 0.7968
No log 10.4255 490 0.7068 0.4096 0.7068 0.8407
No log 10.4681 492 0.6649 0.4670 0.6649 0.8154
No log 10.5106 494 0.5467 0.5918 0.5467 0.7394
No log 10.5532 496 0.5004 0.5831 0.5004 0.7074
No log 10.5957 498 0.5234 0.5362 0.5234 0.7235
0.3457 10.6383 500 0.5043 0.5593 0.5043 0.7101
0.3457 10.6809 502 0.4839 0.6267 0.4839 0.6956
0.3457 10.7234 504 0.5182 0.5904 0.5182 0.7199
0.3457 10.7660 506 0.5483 0.5554 0.5483 0.7405
0.3457 10.8085 508 0.5513 0.5895 0.5513 0.7425
0.3457 10.8511 510 0.5296 0.5702 0.5296 0.7277
0.3457 10.8936 512 0.5215 0.5861 0.5215 0.7222
0.3457 10.9362 514 0.5064 0.5430 0.5064 0.7116
0.3457 10.9787 516 0.5137 0.4742 0.5137 0.7167
0.3457 11.0213 518 0.5091 0.4742 0.5091 0.7135
0.3457 11.0638 520 0.5210 0.5152 0.5210 0.7218
0.3457 11.1064 522 0.5516 0.4802 0.5516 0.7427

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k9_task7_organization

Finetuned
(4019)
this model