ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k8_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8642
  • Qwk: 0.3796
  • Mse: 0.8642
  • Rmse: 0.9296

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0488 2 4.2311 0.0086 4.2311 2.0570
No log 0.0976 4 2.4475 -0.0473 2.4475 1.5645
No log 0.1463 6 1.3700 0.0 1.3700 1.1705
No log 0.1951 8 1.0869 0.2391 1.0869 1.0425
No log 0.2439 10 1.0884 0.2448 1.0884 1.0433
No log 0.2927 12 0.9950 0.1418 0.9950 0.9975
No log 0.3415 14 1.0718 0.1981 1.0718 1.0353
No log 0.3902 16 1.0766 0.2466 1.0766 1.0376
No log 0.4390 18 1.1412 0.2271 1.1412 1.0682
No log 0.4878 20 1.1114 0.2125 1.1114 1.0542
No log 0.5366 22 1.0175 0.2639 1.0175 1.0087
No log 0.5854 24 0.9695 0.2291 0.9695 0.9846
No log 0.6341 26 0.9503 0.2517 0.9503 0.9748
No log 0.6829 28 0.9333 0.2424 0.9333 0.9661
No log 0.7317 30 0.9382 0.3156 0.9382 0.9686
No log 0.7805 32 0.9728 0.2911 0.9728 0.9863
No log 0.8293 34 1.0004 0.2133 1.0004 1.0002
No log 0.8780 36 1.0658 0.2805 1.0658 1.0324
No log 0.9268 38 1.1731 0.2354 1.1731 1.0831
No log 0.9756 40 1.1084 0.2354 1.1084 1.0528
No log 1.0244 42 1.1291 0.2203 1.1291 1.0626
No log 1.0732 44 1.0637 0.2547 1.0637 1.0313
No log 1.1220 46 0.9642 0.2736 0.9642 0.9819
No log 1.1707 48 0.9111 0.3214 0.9111 0.9545
No log 1.2195 50 0.8753 0.3515 0.8753 0.9356
No log 1.2683 52 0.8421 0.3877 0.8421 0.9177
No log 1.3171 54 0.8834 0.4246 0.8834 0.9399
No log 1.3659 56 0.8476 0.4295 0.8476 0.9206
No log 1.4146 58 0.9080 0.2390 0.9080 0.9529
No log 1.4634 60 0.9917 0.2424 0.9917 0.9959
No log 1.5122 62 0.8673 0.3243 0.8673 0.9313
No log 1.5610 64 0.7778 0.4643 0.7778 0.8819
No log 1.6098 66 0.7806 0.3941 0.7806 0.8835
No log 1.6585 68 1.0397 0.3677 1.0397 1.0197
No log 1.7073 70 1.1646 0.2614 1.1646 1.0792
No log 1.7561 72 0.9699 0.3677 0.9699 0.9848
No log 1.8049 74 0.8150 0.4608 0.8150 0.9028
No log 1.8537 76 1.0893 0.3502 1.0893 1.0437
No log 1.9024 78 0.9138 0.4465 0.9138 0.9559
No log 1.9512 80 0.7611 0.4051 0.7611 0.8724
No log 2.0 82 0.7705 0.4051 0.7705 0.8778
No log 2.0488 84 0.8238 0.3853 0.8238 0.9076
No log 2.0976 86 0.8165 0.3795 0.8165 0.9036
No log 2.1463 88 0.8200 0.4745 0.8200 0.9055
No log 2.1951 90 0.9203 0.3959 0.9203 0.9593
No log 2.2439 92 0.9975 0.3944 0.9975 0.9988
No log 2.2927 94 0.9407 0.3861 0.9407 0.9699
No log 2.3415 96 0.9878 0.3960 0.9878 0.9939
No log 2.3902 98 1.0521 0.3519 1.0521 1.0257
No log 2.4390 100 1.0865 0.3622 1.0865 1.0423
No log 2.4878 102 1.0187 0.3846 1.0187 1.0093
No log 2.5366 104 1.0746 0.4053 1.0746 1.0366
No log 2.5854 106 1.0393 0.4067 1.0393 1.0195
No log 2.6341 108 0.8515 0.4869 0.8515 0.9228
No log 2.6829 110 0.8255 0.5557 0.8255 0.9086
No log 2.7317 112 0.7885 0.5304 0.7885 0.8880
No log 2.7805 114 1.0187 0.4050 1.0187 1.0093
No log 2.8293 116 1.0152 0.3832 1.0152 1.0076
No log 2.8780 118 0.7612 0.5318 0.7612 0.8725
No log 2.9268 120 0.8621 0.4346 0.8621 0.9285
No log 2.9756 122 0.8338 0.4435 0.8338 0.9131
No log 3.0244 124 0.8323 0.4590 0.8323 0.9123
No log 3.0732 126 1.1860 0.3937 1.1860 1.0890
No log 3.1220 128 1.1488 0.3937 1.1488 1.0718
No log 3.1707 130 0.8686 0.4214 0.8686 0.9320
No log 3.2195 132 0.8374 0.4321 0.8374 0.9151
No log 3.2683 134 1.0794 0.3424 1.0794 1.0389
No log 3.3171 136 1.0144 0.3458 1.0144 1.0071
No log 3.3659 138 0.9114 0.4526 0.9114 0.9547
No log 3.4146 140 1.1101 0.3344 1.1101 1.0536
No log 3.4634 142 1.1782 0.3226 1.1782 1.0855
No log 3.5122 144 1.0925 0.3463 1.0925 1.0452
No log 3.5610 146 1.0143 0.3154 1.0143 1.0071
No log 3.6098 148 0.9906 0.3571 0.9906 0.9953
No log 3.6585 150 0.9311 0.4368 0.9311 0.9650
No log 3.7073 152 0.9206 0.4483 0.9206 0.9595
No log 3.7561 154 0.8794 0.4478 0.8794 0.9378
No log 3.8049 156 0.8754 0.4478 0.8754 0.9356
No log 3.8537 158 0.8817 0.4388 0.8817 0.9390
No log 3.9024 160 0.8978 0.4388 0.8978 0.9475
No log 3.9512 162 0.9324 0.4599 0.9324 0.9656
No log 4.0 164 0.9247 0.4266 0.9247 0.9616
No log 4.0488 166 0.9426 0.4620 0.9426 0.9709
No log 4.0976 168 0.9230 0.4435 0.9230 0.9607
No log 4.1463 170 0.9774 0.4161 0.9774 0.9886
No log 4.1951 172 1.0108 0.3928 1.0108 1.0054
No log 4.2439 174 0.9053 0.4450 0.9053 0.9515
No log 4.2927 176 0.8925 0.4433 0.8925 0.9447
No log 4.3415 178 0.9344 0.4054 0.9344 0.9666
No log 4.3902 180 0.9143 0.4081 0.9143 0.9562
No log 4.4390 182 0.8589 0.4861 0.8589 0.9268
No log 4.4878 184 0.8602 0.4676 0.8602 0.9275
No log 4.5366 186 0.8856 0.4263 0.8856 0.9410
No log 4.5854 188 0.8884 0.4257 0.8884 0.9426
No log 4.6341 190 0.8760 0.4440 0.8760 0.9360
No log 4.6829 192 0.8565 0.4601 0.8565 0.9255
No log 4.7317 194 0.8521 0.4381 0.8521 0.9231
No log 4.7805 196 0.8930 0.4760 0.8930 0.9450
No log 4.8293 198 0.9640 0.4040 0.9640 0.9818
No log 4.8780 200 0.9500 0.3530 0.9500 0.9747
No log 4.9268 202 0.8772 0.4583 0.8772 0.9366
No log 4.9756 204 0.9316 0.3528 0.9316 0.9652
No log 5.0244 206 0.9181 0.3014 0.9181 0.9582
No log 5.0732 208 0.9444 0.3564 0.9444 0.9718
No log 5.1220 210 0.9316 0.3779 0.9316 0.9652
No log 5.1707 212 0.9523 0.3590 0.9523 0.9759
No log 5.2195 214 0.9949 0.3861 0.9949 0.9975
No log 5.2683 216 0.9644 0.3860 0.9644 0.9821
No log 5.3171 218 1.0257 0.3369 1.0257 1.0128
No log 5.3659 220 1.0828 0.2547 1.0828 1.0406
No log 5.4146 222 0.9498 0.4611 0.9498 0.9746
No log 5.4634 224 0.9377 0.3860 0.9377 0.9683
No log 5.5122 226 1.0308 0.3913 1.0308 1.0153
No log 5.5610 228 1.0623 0.3567 1.0623 1.0307
No log 5.6098 230 0.9870 0.3652 0.9870 0.9935
No log 5.6585 232 0.9844 0.3998 0.9844 0.9922
No log 5.7073 234 0.9793 0.4280 0.9793 0.9896
No log 5.7561 236 1.0169 0.3714 1.0169 1.0084
No log 5.8049 238 1.0262 0.3714 1.0262 1.0130
No log 5.8537 240 0.9763 0.4020 0.9763 0.9881
No log 5.9024 242 0.9526 0.4514 0.9526 0.9760
No log 5.9512 244 1.0001 0.4150 1.0001 1.0001
No log 6.0 246 1.1897 0.3562 1.1897 1.0907
No log 6.0488 248 1.2585 0.3606 1.2585 1.1218
No log 6.0976 250 1.1830 0.3359 1.1830 1.0877
No log 6.1463 252 1.0856 0.3761 1.0856 1.0419
No log 6.1951 254 1.0330 0.4141 1.0330 1.0164
No log 6.2439 256 0.9658 0.4254 0.9658 0.9828
No log 6.2927 258 0.8751 0.3860 0.8751 0.9355
No log 6.3415 260 0.8994 0.4203 0.8994 0.9484
No log 6.3902 262 0.9711 0.3691 0.9711 0.9854
No log 6.4390 264 0.9845 0.3691 0.9845 0.9922
No log 6.4878 266 1.0089 0.3788 1.0089 1.0044
No log 6.5366 268 1.0422 0.3914 1.0422 1.0209
No log 6.5854 270 1.0221 0.4662 1.0221 1.0110
No log 6.6341 272 1.0870 0.3504 1.0870 1.0426
No log 6.6829 274 1.2026 0.3879 1.2026 1.0966
No log 6.7317 276 1.0959 0.3625 1.0959 1.0468
No log 6.7805 278 0.9611 0.4647 0.9611 0.9804
No log 6.8293 280 0.9457 0.3771 0.9457 0.9725
No log 6.8780 282 0.8656 0.4404 0.8656 0.9304
No log 6.9268 284 0.8951 0.5085 0.8951 0.9461
No log 6.9756 286 1.2060 0.3424 1.2060 1.0982
No log 7.0244 288 1.2448 0.3439 1.2448 1.1157
No log 7.0732 290 1.0112 0.3741 1.0112 1.0056
No log 7.1220 292 0.8422 0.4304 0.8422 0.9177
No log 7.1707 294 0.8783 0.3973 0.8783 0.9372
No log 7.2195 296 1.0728 0.3200 1.0728 1.0357
No log 7.2683 298 1.1560 0.3502 1.1560 1.0752
No log 7.3171 300 1.1420 0.3657 1.1420 1.0686
No log 7.3659 302 1.1312 0.3490 1.1312 1.0636
No log 7.4146 304 1.1196 0.3820 1.1196 1.0581
No log 7.4634 306 1.0398 0.3998 1.0398 1.0197
No log 7.5122 308 0.9737 0.3285 0.9737 0.9868
No log 7.5610 310 0.9487 0.2887 0.9487 0.9740
No log 7.6098 312 0.9243 0.4234 0.9243 0.9614
No log 7.6585 314 0.9702 0.3544 0.9702 0.9850
No log 7.7073 316 1.0008 0.3855 1.0008 1.0004
No log 7.7561 318 1.0107 0.4079 1.0107 1.0053
No log 7.8049 320 1.0331 0.4390 1.0331 1.0164
No log 7.8537 322 1.0174 0.3689 1.0174 1.0087
No log 7.9024 324 1.0308 0.3662 1.0308 1.0153
No log 7.9512 326 1.0107 0.3380 1.0107 1.0053
No log 8.0 328 0.9527 0.3474 0.9527 0.9760
No log 8.0488 330 0.9318 0.3188 0.9318 0.9653
No log 8.0976 332 0.9172 0.3474 0.9172 0.9577
No log 8.1463 334 0.9261 0.3957 0.9261 0.9624
No log 8.1951 336 0.9521 0.3558 0.9521 0.9757
No log 8.2439 338 0.9038 0.3691 0.9038 0.9507
No log 8.2927 340 0.8006 0.4227 0.8006 0.8948
No log 8.3415 342 0.7615 0.4629 0.7615 0.8726
No log 8.3902 344 0.7796 0.4727 0.7796 0.8830
No log 8.4390 346 0.8310 0.4823 0.8310 0.9116
No log 8.4878 348 0.9095 0.4369 0.9095 0.9537
No log 8.5366 350 0.9345 0.4468 0.9345 0.9667
No log 8.5854 352 0.8291 0.4911 0.8291 0.9106
No log 8.6341 354 0.7778 0.5266 0.7778 0.8819
No log 8.6829 356 0.7341 0.5070 0.7341 0.8568
No log 8.7317 358 0.7195 0.5195 0.7195 0.8482
No log 8.7805 360 0.7142 0.5669 0.7142 0.8451
No log 8.8293 362 0.7703 0.4536 0.7703 0.8776
No log 8.8780 364 0.7523 0.5854 0.7523 0.8674
No log 8.9268 366 0.7626 0.5528 0.7626 0.8733
No log 8.9756 368 0.8558 0.4735 0.8558 0.9251
No log 9.0244 370 0.9430 0.4500 0.9430 0.9711
No log 9.0732 372 0.8616 0.4723 0.8616 0.9282
No log 9.1220 374 0.8108 0.4772 0.8108 0.9004
No log 9.1707 376 0.8255 0.4828 0.8255 0.9086
No log 9.2195 378 0.8670 0.4860 0.8670 0.9311
No log 9.2683 380 0.9325 0.4070 0.9325 0.9657
No log 9.3171 382 0.8743 0.4482 0.8743 0.9350
No log 9.3659 384 0.8241 0.4378 0.8241 0.9078
No log 9.4146 386 0.8627 0.4075 0.8627 0.9288
No log 9.4634 388 0.8798 0.4271 0.8798 0.9380
No log 9.5122 390 0.9058 0.4550 0.9058 0.9517
No log 9.5610 392 0.9282 0.4015 0.9282 0.9635
No log 9.6098 394 0.8946 0.4774 0.8946 0.9459
No log 9.6585 396 0.8841 0.4352 0.8841 0.9403
No log 9.7073 398 0.8913 0.3668 0.8913 0.9441
No log 9.7561 400 0.8672 0.4461 0.8672 0.9312
No log 9.8049 402 0.7972 0.4244 0.7972 0.8929
No log 9.8537 404 0.7685 0.4691 0.7685 0.8766
No log 9.9024 406 0.7617 0.4816 0.7617 0.8727
No log 9.9512 408 0.7737 0.5391 0.7737 0.8796
No log 10.0 410 0.8147 0.5192 0.8147 0.9026
No log 10.0488 412 0.8519 0.4430 0.8519 0.9230
No log 10.0976 414 0.8520 0.5474 0.8520 0.9230
No log 10.1463 416 0.8423 0.5186 0.8423 0.9178
No log 10.1951 418 0.8220 0.4186 0.8220 0.9066
No log 10.2439 420 0.8251 0.4056 0.8251 0.9084
No log 10.2927 422 0.8659 0.4023 0.8659 0.9305
No log 10.3415 424 0.8711 0.3824 0.8711 0.9333
No log 10.3902 426 0.8676 0.3346 0.8676 0.9314
No log 10.4390 428 0.9141 0.3956 0.9141 0.9561
No log 10.4878 430 0.9442 0.4225 0.9442 0.9717
No log 10.5366 432 0.9416 0.4053 0.9416 0.9704
No log 10.5854 434 0.9167 0.3674 0.9167 0.9575
No log 10.6341 436 0.8645 0.3875 0.8645 0.9298
No log 10.6829 438 0.8447 0.3875 0.8447 0.9191
No log 10.7317 440 0.8463 0.4630 0.8463 0.9199
No log 10.7805 442 0.8400 0.4981 0.8400 0.9165
No log 10.8293 444 0.8828 0.3973 0.8828 0.9396
No log 10.8780 446 0.9147 0.3844 0.9147 0.9564
No log 10.9268 448 0.8896 0.4241 0.8896 0.9432
No log 10.9756 450 0.8654 0.4974 0.8654 0.9303
No log 11.0244 452 0.8973 0.4958 0.8973 0.9473
No log 11.0732 454 0.9796 0.4910 0.9796 0.9897
No log 11.1220 456 1.0782 0.4181 1.0782 1.0384
No log 11.1707 458 1.0695 0.3956 1.0695 1.0342
No log 11.2195 460 0.9139 0.4119 0.9139 0.9560
No log 11.2683 462 0.7878 0.4878 0.7878 0.8876
No log 11.3171 464 0.7708 0.5432 0.7708 0.8780
No log 11.3659 466 0.7659 0.5054 0.7659 0.8751
No log 11.4146 468 0.7714 0.4511 0.7714 0.8783
No log 11.4634 470 0.7939 0.5195 0.7939 0.8910
No log 11.5122 472 0.8390 0.4710 0.8390 0.9159
No log 11.5610 474 0.8356 0.4948 0.8356 0.9141
No log 11.6098 476 0.8286 0.4853 0.8286 0.9103
No log 11.6585 478 0.8674 0.5175 0.8674 0.9313
No log 11.7073 480 0.8594 0.4853 0.8594 0.9270
No log 11.7561 482 0.8701 0.4853 0.8701 0.9328
No log 11.8049 484 0.8499 0.4975 0.8499 0.9219
No log 11.8537 486 0.8340 0.4329 0.8340 0.9132
No log 11.9024 488 0.8360 0.3785 0.8360 0.9143
No log 11.9512 490 0.8435 0.3785 0.8435 0.9184
No log 12.0 492 0.8227 0.4391 0.8227 0.9070
No log 12.0488 494 0.8701 0.4843 0.8701 0.9328
No log 12.0976 496 0.9278 0.4606 0.9278 0.9632
No log 12.1463 498 0.9215 0.4606 0.9215 0.9600
0.3411 12.1951 500 0.9111 0.4606 0.9111 0.9545
0.3411 12.2439 502 0.8969 0.4480 0.8969 0.9470
0.3411 12.2927 504 0.8575 0.4359 0.8575 0.9260
0.3411 12.3415 506 0.8275 0.4615 0.8275 0.9097
0.3411 12.3902 508 0.8229 0.4391 0.8229 0.9071
0.3411 12.4390 510 0.8306 0.4391 0.8306 0.9114
0.3411 12.4878 512 0.8234 0.4974 0.8234 0.9074
0.3411 12.5366 514 0.8046 0.4974 0.8046 0.8970
0.3411 12.5854 516 0.7741 0.4757 0.7741 0.8798
0.3411 12.6341 518 0.7954 0.4560 0.7954 0.8919
0.3411 12.6829 520 0.8074 0.4560 0.8074 0.8985
0.3411 12.7317 522 0.7765 0.4757 0.7765 0.8812
0.3411 12.7805 524 0.8065 0.4599 0.8065 0.8981
0.3411 12.8293 526 0.8086 0.4343 0.8086 0.8992
0.3411 12.8780 528 0.7710 0.4727 0.7710 0.8781
0.3411 12.9268 530 0.7733 0.4727 0.7733 0.8794
0.3411 12.9756 532 0.8041 0.4471 0.8041 0.8967
0.3411 13.0244 534 0.8530 0.3738 0.8530 0.9236
0.3411 13.0732 536 0.8398 0.3863 0.8398 0.9164
0.3411 13.1220 538 0.8078 0.4071 0.8078 0.8988
0.3411 13.1707 540 0.7872 0.4858 0.7872 0.8872
0.3411 13.2195 542 0.8072 0.3976 0.8072 0.8984
0.3411 13.2683 544 0.8137 0.3706 0.8137 0.9020
0.3411 13.3171 546 0.8061 0.3804 0.8061 0.8978
0.3411 13.3659 548 0.8348 0.4455 0.8348 0.9137
0.3411 13.4146 550 0.8126 0.4847 0.8126 0.9015
0.3411 13.4634 552 0.8019 0.4778 0.8019 0.8955
0.3411 13.5122 554 0.8001 0.4810 0.8001 0.8945
0.3411 13.5610 556 0.7895 0.4675 0.7895 0.8886
0.3411 13.6098 558 0.7926 0.4938 0.7926 0.8903
0.3411 13.6585 560 0.7876 0.5135 0.7876 0.8875
0.3411 13.7073 562 0.8371 0.4723 0.8371 0.9149
0.3411 13.7561 564 0.9934 0.4354 0.9934 0.9967
0.3411 13.8049 566 1.1321 0.3353 1.1321 1.0640
0.3411 13.8537 568 1.0762 0.3721 1.0762 1.0374
0.3411 13.9024 570 0.9104 0.3672 0.9104 0.9541
0.3411 13.9512 572 0.7507 0.5543 0.7507 0.8664
0.3411 14.0 574 0.7321 0.5214 0.7321 0.8556
0.3411 14.0488 576 0.7668 0.4624 0.7668 0.8757
0.3411 14.0976 578 0.7679 0.4983 0.7679 0.8763
0.3411 14.1463 580 0.7732 0.5669 0.7732 0.8793
0.3411 14.1951 582 0.7705 0.5602 0.7705 0.8778
0.3411 14.2439 584 0.7693 0.4962 0.7693 0.8771
0.3411 14.2927 586 0.7704 0.4962 0.7704 0.8778
0.3411 14.3415 588 0.7691 0.5304 0.7691 0.8770
0.3411 14.3902 590 0.7646 0.5446 0.7646 0.8744
0.3411 14.4390 592 0.7665 0.5343 0.7665 0.8755
0.3411 14.4878 594 0.7657 0.5330 0.7657 0.8751
0.3411 14.5366 596 0.7466 0.5463 0.7466 0.8641
0.3411 14.5854 598 0.7419 0.5463 0.7419 0.8613
0.3411 14.6341 600 0.7353 0.5463 0.7353 0.8575
0.3411 14.6829 602 0.7399 0.6466 0.7399 0.8602
0.3411 14.7317 604 0.7578 0.6024 0.7578 0.8705
0.3411 14.7805 606 0.7586 0.4772 0.7586 0.8710
0.3411 14.8293 608 0.7720 0.4273 0.7720 0.8787
0.3411 14.8780 610 0.7894 0.4511 0.7894 0.8885
0.3411 14.9268 612 0.7963 0.4511 0.7963 0.8924
0.3411 14.9756 614 0.7930 0.4511 0.7930 0.8905
0.3411 15.0244 616 0.8108 0.4646 0.8108 0.9005
0.3411 15.0732 618 0.8246 0.4328 0.8246 0.9080
0.3411 15.1220 620 0.8218 0.5112 0.8218 0.9065
0.3411 15.1707 622 0.8233 0.4398 0.8233 0.9074
0.3411 15.2195 624 0.8263 0.4008 0.8263 0.9090
0.3411 15.2683 626 0.7983 0.4511 0.7983 0.8935
0.3411 15.3171 628 0.8269 0.4513 0.8269 0.9093
0.3411 15.3659 630 0.8642 0.3796 0.8642 0.9296

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k8_task5_organization

Finetuned
(4019)
this model