ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k17_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9309
  • Qwk: 0.3483
  • Mse: 0.9309
  • Rmse: 0.9648

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0370 2 4.0746 0.0118 4.0746 2.0186
No log 0.0741 4 2.3197 0.1122 2.3197 1.5231
No log 0.1111 6 1.5419 0.0294 1.5419 1.2417
No log 0.1481 8 1.6033 0.0416 1.6033 1.2662
No log 0.1852 10 1.2339 0.1460 1.2339 1.1108
No log 0.2222 12 1.0530 0.2787 1.0530 1.0262
No log 0.2593 14 1.1355 0.1738 1.1355 1.0656
No log 0.2963 16 1.2838 0.0 1.2838 1.1330
No log 0.3333 18 1.3371 0.0399 1.3371 1.1563
No log 0.3704 20 1.2726 0.0496 1.2726 1.1281
No log 0.4074 22 1.0804 0.2150 1.0804 1.0394
No log 0.4444 24 0.9689 0.3139 0.9689 0.9843
No log 0.4815 26 0.9121 0.3795 0.9121 0.9550
No log 0.5185 28 1.1591 0.1322 1.1591 1.0766
No log 0.5556 30 1.2844 0.1187 1.2844 1.1333
No log 0.5926 32 1.0284 0.2513 1.0284 1.0141
No log 0.6296 34 0.9979 0.3184 0.9979 0.9989
No log 0.6667 36 1.0641 0.2100 1.0641 1.0316
No log 0.7037 38 1.2277 0.1187 1.2277 1.1080
No log 0.7407 40 1.1566 0.1460 1.1566 1.0754
No log 0.7778 42 1.1112 0.2074 1.1112 1.0541
No log 0.8148 44 1.0997 0.2441 1.0997 1.0486
No log 0.8519 46 1.0647 0.2049 1.0647 1.0319
No log 0.8889 48 1.0423 0.2293 1.0423 1.0209
No log 0.9259 50 1.0636 0.2687 1.0636 1.0313
No log 0.9630 52 1.2205 0.1991 1.2205 1.1048
No log 1.0 54 1.2706 0.2133 1.2706 1.1272
No log 1.0370 56 1.2094 0.2808 1.2094 1.0997
No log 1.0741 58 1.0683 0.3188 1.0683 1.0336
No log 1.1111 60 1.0754 0.1962 1.0754 1.0370
No log 1.1481 62 1.0609 0.2365 1.0609 1.0300
No log 1.1852 64 1.0242 0.3133 1.0242 1.0120
No log 1.2222 66 1.0905 0.3329 1.0905 1.0443
No log 1.2593 68 1.0516 0.3084 1.0516 1.0255
No log 1.2963 70 1.0082 0.3172 1.0082 1.0041
No log 1.3333 72 1.0786 0.1532 1.0786 1.0385
No log 1.3704 74 1.0393 0.2291 1.0393 1.0194
No log 1.4074 76 1.0585 0.3066 1.0585 1.0288
No log 1.4444 78 1.1233 0.2686 1.1233 1.0599
No log 1.4815 80 1.0438 0.2507 1.0438 1.0217
No log 1.5185 82 1.0178 0.2114 1.0178 1.0089
No log 1.5556 84 1.0044 0.2618 1.0044 1.0022
No log 1.5926 86 0.9474 0.2572 0.9474 0.9733
No log 1.6296 88 0.9165 0.3097 0.9165 0.9574
No log 1.6667 90 0.9098 0.3230 0.9098 0.9538
No log 1.7037 92 0.9408 0.3070 0.9408 0.9699
No log 1.7407 94 1.0572 0.3643 1.0572 1.0282
No log 1.7778 96 0.9826 0.3067 0.9826 0.9913
No log 1.8148 98 0.9179 0.3030 0.9179 0.9581
No log 1.8519 100 0.8791 0.3229 0.8791 0.9376
No log 1.8889 102 0.9671 0.4089 0.9671 0.9834
No log 1.9259 104 0.9843 0.4089 0.9843 0.9921
No log 1.9630 106 1.0034 0.4089 1.0034 1.0017
No log 2.0 108 1.0185 0.4588 1.0185 1.0092
No log 2.0370 110 0.9960 0.4823 0.9960 0.9980
No log 2.0741 112 1.0001 0.4620 1.0001 1.0001
No log 2.1111 114 0.9592 0.4102 0.9592 0.9794
No log 2.1481 116 1.0191 0.4268 1.0191 1.0095
No log 2.1852 118 0.9768 0.3448 0.9768 0.9883
No log 2.2222 120 1.1717 0.3935 1.1717 1.0824
No log 2.2593 122 1.2510 0.3445 1.2510 1.1185
No log 2.2963 124 1.0306 0.4334 1.0306 1.0152
No log 2.3333 126 0.9067 0.3627 0.9067 0.9522
No log 2.3704 128 1.0391 0.4288 1.0391 1.0194
No log 2.4074 130 0.9477 0.2986 0.9477 0.9735
No log 2.4444 132 0.8563 0.3693 0.8563 0.9253
No log 2.4815 134 0.9392 0.3229 0.9392 0.9691
No log 2.5185 136 0.9425 0.4119 0.9425 0.9708
No log 2.5556 138 0.9395 0.3711 0.9395 0.9693
No log 2.5926 140 0.9872 0.3983 0.9872 0.9936
No log 2.6296 142 1.0200 0.3590 1.0200 1.0100
No log 2.6667 144 1.1855 0.4096 1.1855 1.0888
No log 2.7037 146 1.2072 0.3723 1.2072 1.0987
No log 2.7407 148 1.1432 0.3820 1.1432 1.0692
No log 2.7778 150 1.1854 0.3484 1.1854 1.0888
No log 2.8148 152 1.1163 0.3167 1.1163 1.0566
No log 2.8519 154 1.2019 0.3552 1.2019 1.0963
No log 2.8889 156 1.4508 0.2567 1.4508 1.2045
No log 2.9259 158 1.3139 0.3015 1.3139 1.1463
No log 2.9630 160 1.0327 0.2871 1.0327 1.0162
No log 3.0 162 1.0025 0.4019 1.0025 1.0012
No log 3.0370 164 0.9578 0.3457 0.9578 0.9787
No log 3.0741 166 1.0294 0.3710 1.0294 1.0146
No log 3.1111 168 1.0655 0.4218 1.0655 1.0322
No log 3.1481 170 1.0051 0.3383 1.0051 1.0025
No log 3.1852 172 0.9176 0.3151 0.9176 0.9579
No log 3.2222 174 0.9458 0.3651 0.9458 0.9725
No log 3.2593 176 0.9462 0.3522 0.9462 0.9727
No log 3.2963 178 1.0321 0.3806 1.0321 1.0159
No log 3.3333 180 1.1424 0.3961 1.1424 1.0688
No log 3.3704 182 1.1758 0.3645 1.1758 1.0843
No log 3.4074 184 1.1884 0.3640 1.1884 1.0902
No log 3.4444 186 1.1823 0.4016 1.1823 1.0873
No log 3.4815 188 1.0982 0.3498 1.0982 1.0480
No log 3.5185 190 1.0641 0.4037 1.0641 1.0315
No log 3.5556 192 1.0609 0.4343 1.0609 1.0300
No log 3.5926 194 1.0674 0.4242 1.0674 1.0331
No log 3.6296 196 1.1112 0.4060 1.1112 1.0541
No log 3.6667 198 1.1364 0.4276 1.1364 1.0660
No log 3.7037 200 1.0660 0.4611 1.0660 1.0325
No log 3.7407 202 1.0340 0.4996 1.0340 1.0169
No log 3.7778 204 0.9969 0.4996 0.9969 0.9985
No log 3.8148 206 0.9657 0.3872 0.9657 0.9827
No log 3.8519 208 0.9498 0.3856 0.9498 0.9746
No log 3.8889 210 0.9267 0.4011 0.9267 0.9627
No log 3.9259 212 0.8994 0.4015 0.8994 0.9484
No log 3.9630 214 0.8932 0.3726 0.8932 0.9451
No log 4.0 216 0.9093 0.4002 0.9093 0.9536
No log 4.0370 218 0.9369 0.4331 0.9369 0.9679
No log 4.0741 220 0.9130 0.4217 0.9130 0.9555
No log 4.1111 222 1.0073 0.5050 1.0073 1.0036
No log 4.1481 224 1.0398 0.5222 1.0398 1.0197
No log 4.1852 226 0.9383 0.4532 0.9383 0.9687
No log 4.2222 228 0.9296 0.4410 0.9296 0.9642
No log 4.2593 230 0.9458 0.3621 0.9458 0.9725
No log 4.2963 232 0.8672 0.3725 0.8672 0.9312
No log 4.3333 234 0.8420 0.3878 0.8420 0.9176
No log 4.3704 236 0.8585 0.3682 0.8585 0.9266
No log 4.4074 238 0.8869 0.3861 0.8869 0.9418
No log 4.4444 240 0.8617 0.4273 0.8617 0.9283
No log 4.4815 242 0.8577 0.3980 0.8577 0.9261
No log 4.5185 244 0.9132 0.5255 0.9132 0.9556
No log 4.5556 246 0.9462 0.5255 0.9462 0.9727
No log 4.5926 248 0.9046 0.3977 0.9046 0.9511
No log 4.6296 250 0.9324 0.4221 0.9324 0.9656
No log 4.6667 252 0.9751 0.4321 0.9751 0.9875
No log 4.7037 254 1.0307 0.4736 1.0307 1.0152
No log 4.7407 256 1.1550 0.4486 1.1550 1.0747
No log 4.7778 258 1.1284 0.4205 1.1284 1.0623
No log 4.8148 260 1.0502 0.3715 1.0502 1.0248
No log 4.8519 262 1.0288 0.3500 1.0288 1.0143
No log 4.8889 264 1.0234 0.3786 1.0234 1.0116
No log 4.9259 266 1.0291 0.3680 1.0291 1.0144
No log 4.9630 268 1.0502 0.4386 1.0502 1.0248
No log 5.0 270 1.0151 0.3946 1.0151 1.0075
No log 5.0370 272 0.9856 0.3464 0.9856 0.9928
No log 5.0741 274 1.0192 0.3680 1.0192 1.0096
No log 5.1111 276 1.1067 0.3863 1.1067 1.0520
No log 5.1481 278 1.1648 0.3679 1.1648 1.0793
No log 5.1852 280 1.1177 0.3813 1.1177 1.0572
No log 5.2222 282 1.0290 0.4065 1.0290 1.0144
No log 5.2593 284 1.0183 0.4436 1.0183 1.0091
No log 5.2963 286 0.9893 0.4052 0.9893 0.9946
No log 5.3333 288 0.9790 0.4412 0.9790 0.9894
No log 5.3704 290 0.9958 0.3933 0.9958 0.9979
No log 5.4074 292 1.0058 0.3711 1.0058 1.0029
No log 5.4444 294 1.0369 0.4349 1.0369 1.0183
No log 5.4815 296 1.0283 0.4349 1.0283 1.0141
No log 5.5185 298 0.9429 0.4555 0.9429 0.9711
No log 5.5556 300 0.9226 0.4449 0.9226 0.9605
No log 5.5926 302 0.9758 0.4635 0.9758 0.9878
No log 5.6296 304 0.9686 0.4567 0.9686 0.9842
No log 5.6667 306 0.9656 0.4859 0.9656 0.9826
No log 5.7037 308 1.0548 0.4815 1.0548 1.0270
No log 5.7407 310 1.0313 0.4625 1.0313 1.0156
No log 5.7778 312 0.9753 0.4552 0.9753 0.9876
No log 5.8148 314 1.0088 0.4426 1.0088 1.0044
No log 5.8519 316 0.9783 0.4376 0.9783 0.9891
No log 5.8889 318 0.9939 0.4062 0.9939 0.9969
No log 5.9259 320 1.0814 0.5029 1.0814 1.0399
No log 5.9630 322 1.1790 0.4642 1.1790 1.0858
No log 6.0 324 1.1101 0.4359 1.1101 1.0536
No log 6.0370 326 1.0175 0.4689 1.0175 1.0087
No log 6.0741 328 1.0040 0.4547 1.0040 1.0020
No log 6.1111 330 0.9971 0.4659 0.9971 0.9986
No log 6.1481 332 0.9715 0.4301 0.9715 0.9857
No log 6.1852 334 0.9505 0.4366 0.9505 0.9749
No log 6.2222 336 0.9549 0.4479 0.9549 0.9772
No log 6.2593 338 0.9638 0.4484 0.9638 0.9817
No log 6.2963 340 0.9786 0.4370 0.9786 0.9892
No log 6.3333 342 0.9663 0.4370 0.9663 0.9830
No log 6.3704 344 0.9742 0.4079 0.9742 0.9870
No log 6.4074 346 0.9457 0.3747 0.9457 0.9725
No log 6.4444 348 0.8702 0.4141 0.8702 0.9328
No log 6.4815 350 0.8650 0.4033 0.8650 0.9300
No log 6.5185 352 0.8643 0.3740 0.8643 0.9297
No log 6.5556 354 0.9158 0.3634 0.9158 0.9570
No log 6.5926 356 1.0128 0.3341 1.0128 1.0064
No log 6.6296 358 1.0427 0.4403 1.0427 1.0211
No log 6.6667 360 1.0292 0.4720 1.0292 1.0145
No log 6.7037 362 1.0120 0.4625 1.0120 1.0060
No log 6.7407 364 0.9786 0.4527 0.9786 0.9892
No log 6.7778 366 0.9787 0.4733 0.9787 0.9893
No log 6.8148 368 0.9569 0.4507 0.9569 0.9782
No log 6.8519 370 0.9107 0.4726 0.9107 0.9543
No log 6.8889 372 0.8702 0.4540 0.8702 0.9328
No log 6.9259 374 0.8847 0.4235 0.8847 0.9406
No log 6.9630 376 0.8939 0.4345 0.8939 0.9455
No log 7.0 378 0.9275 0.4898 0.9275 0.9631
No log 7.0370 380 1.0023 0.5182 1.0023 1.0011
No log 7.0741 382 1.0948 0.4278 1.0948 1.0463
No log 7.1111 384 1.1523 0.3759 1.1523 1.0734
No log 7.1481 386 1.0680 0.4681 1.0680 1.0334
No log 7.1852 388 0.9230 0.4401 0.9230 0.9607
No log 7.2222 390 0.8741 0.5427 0.8741 0.9349
No log 7.2593 392 0.8885 0.5427 0.8885 0.9426
No log 7.2963 394 0.8962 0.4853 0.8962 0.9467
No log 7.3333 396 1.0038 0.4482 1.0038 1.0019
No log 7.3704 398 1.0690 0.4585 1.0690 1.0339
No log 7.4074 400 1.0528 0.4486 1.0528 1.0261
No log 7.4444 402 0.9702 0.4945 0.9702 0.9850
No log 7.4815 404 0.9599 0.4972 0.9599 0.9798
No log 7.5185 406 0.9387 0.4406 0.9387 0.9689
No log 7.5556 408 0.9619 0.4048 0.9619 0.9807
No log 7.5926 410 0.9625 0.4392 0.9625 0.9811
No log 7.6296 412 0.9065 0.4881 0.9065 0.9521
No log 7.6667 414 0.9114 0.5347 0.9114 0.9547
No log 7.7037 416 0.9067 0.4950 0.9067 0.9522
No log 7.7407 418 0.8895 0.5385 0.8895 0.9431
No log 7.7778 420 0.9163 0.5660 0.9163 0.9572
No log 7.8148 422 0.8909 0.5489 0.8909 0.9439
No log 7.8519 424 0.8838 0.4902 0.8838 0.9401
No log 7.8889 426 0.9031 0.4911 0.9031 0.9503
No log 7.9259 428 0.9059 0.4907 0.9059 0.9518
No log 7.9630 430 0.9093 0.4853 0.9093 0.9535
No log 8.0 432 0.8876 0.4526 0.8876 0.9421
No log 8.0370 434 0.8686 0.4889 0.8686 0.9320
No log 8.0741 436 0.8951 0.4556 0.8951 0.9461
No log 8.1111 438 1.0183 0.4021 1.0183 1.0091
No log 8.1481 440 0.9824 0.4252 0.9824 0.9911
No log 8.1852 442 0.9068 0.4632 0.9068 0.9523
No log 8.2222 444 0.9011 0.4996 0.9011 0.9493
No log 8.2593 446 0.9474 0.5473 0.9474 0.9733
No log 8.2963 448 0.9479 0.5267 0.9479 0.9736
No log 8.3333 450 0.9792 0.4940 0.9792 0.9895
No log 8.3704 452 1.0485 0.4395 1.0485 1.0240
No log 8.4074 454 1.0080 0.4517 1.0080 1.0040
No log 8.4444 456 0.9831 0.4703 0.9831 0.9915
No log 8.4815 458 1.0717 0.4479 1.0717 1.0352
No log 8.5185 460 1.2055 0.4331 1.2055 1.0980
No log 8.5556 462 1.1988 0.4337 1.1988 1.0949
No log 8.5926 464 1.1413 0.4164 1.1413 1.0683
No log 8.6296 466 1.1169 0.4203 1.1169 1.0568
No log 8.6667 468 1.1169 0.3757 1.1169 1.0568
No log 8.7037 470 1.0860 0.4694 1.0860 1.0421
No log 8.7407 472 1.0022 0.4276 1.0022 1.0011
No log 8.7778 474 0.9250 0.4106 0.9250 0.9618
No log 8.8148 476 0.8820 0.4774 0.8820 0.9392
No log 8.8519 478 0.8976 0.4321 0.8976 0.9474
No log 8.8889 480 0.9255 0.5103 0.9255 0.9620
No log 8.9259 482 1.0502 0.4597 1.0502 1.0248
No log 8.9630 484 1.0882 0.4405 1.0882 1.0432
No log 9.0 486 1.0868 0.4405 1.0868 1.0425
No log 9.0370 488 1.0273 0.4719 1.0273 1.0136
No log 9.0741 490 1.0485 0.4400 1.0485 1.0240
No log 9.1111 492 1.0108 0.4805 1.0108 1.0054
No log 9.1481 494 0.9045 0.4708 0.9045 0.9510
No log 9.1852 496 0.8228 0.4861 0.8228 0.9071
No log 9.2222 498 0.7998 0.4534 0.7998 0.8943
0.3011 9.2593 500 0.8210 0.4861 0.8210 0.9061
0.3011 9.2963 502 0.9193 0.4812 0.9193 0.9588
0.3011 9.3333 504 1.0222 0.4907 1.0222 1.0111
0.3011 9.3704 506 1.0711 0.4601 1.0711 1.0350
0.3011 9.4074 508 0.9715 0.4719 0.9715 0.9856
0.3011 9.4444 510 0.9065 0.4449 0.9065 0.9521
0.3011 9.4815 512 0.8842 0.4664 0.8842 0.9403
0.3011 9.5185 514 0.9183 0.4938 0.9183 0.9583
0.3011 9.5556 516 0.9668 0.4918 0.9668 0.9833
0.3011 9.5926 518 1.0549 0.4397 1.0549 1.0271
0.3011 9.6296 520 1.0271 0.4878 1.0271 1.0134
0.3011 9.6667 522 0.9573 0.4694 0.9573 0.9784
0.3011 9.7037 524 0.8998 0.4266 0.8998 0.9486
0.3011 9.7407 526 0.8987 0.4250 0.8987 0.9480
0.3011 9.7778 528 0.9180 0.3990 0.9180 0.9581
0.3011 9.8148 530 0.9309 0.3483 0.9309 0.9648

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k17_task5_organization

Finetuned
(4019)
this model