ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k5_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8523
  • Qwk: 0.6619
  • Mse: 0.8523
  • Rmse: 0.9232

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0870 2 6.9008 -0.0056 6.9008 2.6269
No log 0.1739 4 4.4127 0.0769 4.4127 2.1006
No log 0.2609 6 2.9803 0.0848 2.9803 1.7264
No log 0.3478 8 2.4008 0.1143 2.4008 1.5495
No log 0.4348 10 1.9131 0.1416 1.9131 1.3831
No log 0.5217 12 1.7442 0.1869 1.7442 1.3207
No log 0.6087 14 1.5677 0.1905 1.5677 1.2521
No log 0.6957 16 1.5796 0.1905 1.5796 1.2568
No log 0.7826 18 1.6192 0.25 1.6192 1.2725
No log 0.8696 20 1.9001 0.3101 1.9001 1.3785
No log 0.9565 22 1.9889 0.3008 1.9889 1.4103
No log 1.0435 24 1.8665 0.3438 1.8665 1.3662
No log 1.1304 26 1.7941 0.4094 1.7941 1.3394
No log 1.2174 28 1.8712 0.3622 1.8712 1.3679
No log 1.3043 30 1.8699 0.368 1.8699 1.3674
No log 1.3913 32 1.8485 0.3415 1.8485 1.3596
No log 1.4783 34 1.8259 0.3471 1.8259 1.3512
No log 1.5652 36 1.7262 0.3710 1.7262 1.3138
No log 1.6522 38 1.7298 0.3846 1.7298 1.3152
No log 1.7391 40 1.8969 0.2963 1.8969 1.3773
No log 1.8261 42 2.1767 0.1560 2.1767 1.4754
No log 1.9130 44 2.0625 0.2014 2.0625 1.4361
No log 2.0 46 1.5943 0.3969 1.5943 1.2626
No log 2.0870 48 1.4335 0.4194 1.4335 1.1973
No log 2.1739 50 1.5933 0.3415 1.5933 1.2622
No log 2.2609 52 1.6332 0.384 1.6332 1.2780
No log 2.3478 54 1.5022 0.3968 1.5022 1.2256
No log 2.4348 56 1.3750 0.4320 1.3750 1.1726
No log 2.5217 58 1.3106 0.4640 1.3106 1.1448
No log 2.6087 60 1.3291 0.4724 1.3291 1.1529
No log 2.6957 62 1.5220 0.3622 1.5220 1.2337
No log 2.7826 64 1.8173 0.2047 1.8173 1.3481
No log 2.8696 66 1.7074 0.2462 1.7074 1.3067
No log 2.9565 68 1.5567 0.368 1.5567 1.2477
No log 3.0435 70 1.4787 0.4211 1.4787 1.2160
No log 3.1304 72 1.4756 0.4638 1.4756 1.2147
No log 3.2174 74 1.4010 0.4211 1.4010 1.1836
No log 3.3043 76 1.3314 0.4603 1.3314 1.1539
No log 3.3913 78 1.4970 0.3651 1.4970 1.2235
No log 3.4783 80 1.1301 0.4882 1.1301 1.0631
No log 3.5652 82 1.0712 0.5271 1.0712 1.0350
No log 3.6522 84 1.2014 0.5564 1.2014 1.0961
No log 3.7391 86 0.9924 0.5455 0.9924 0.9962
No log 3.8261 88 0.9016 0.6308 0.9016 0.9495
No log 3.9130 90 0.8862 0.6466 0.8862 0.9414
No log 4.0 92 0.9742 0.6107 0.9742 0.9870
No log 4.0870 94 1.0809 0.496 1.0809 1.0396
No log 4.1739 96 0.9013 0.6567 0.9013 0.9494
No log 4.2609 98 0.8699 0.6815 0.8699 0.9327
No log 4.3478 100 0.9549 0.6471 0.9549 0.9772
No log 4.4348 102 1.1288 0.5538 1.1288 1.0624
No log 4.5217 104 1.2256 0.5312 1.2256 1.1071
No log 4.6087 106 1.0158 0.5942 1.0158 1.0079
No log 4.6957 108 0.8584 0.7034 0.8584 0.9265
No log 4.7826 110 0.7776 0.7183 0.7776 0.8818
No log 4.8696 112 0.8330 0.6901 0.8330 0.9127
No log 4.9565 114 0.8635 0.6569 0.8635 0.9292
No log 5.0435 116 0.7677 0.6912 0.7677 0.8762
No log 5.1304 118 0.7535 0.7050 0.7535 0.8680
No log 5.2174 120 0.7875 0.6716 0.7875 0.8874
No log 5.3043 122 0.9084 0.6412 0.9084 0.9531
No log 5.3913 124 0.8554 0.6667 0.8554 0.9249
No log 5.4783 126 0.8165 0.6963 0.8165 0.9036
No log 5.5652 128 0.8159 0.7059 0.8159 0.9033
No log 5.6522 130 0.8503 0.6767 0.8503 0.9221
No log 5.7391 132 0.9596 0.6412 0.9596 0.9796
No log 5.8261 134 0.8851 0.6917 0.8851 0.9408
No log 5.9130 136 1.1075 0.5874 1.1075 1.0524
No log 6.0 138 1.7224 0.4886 1.7224 1.3124
No log 6.0870 140 2.0422 0.4041 2.0422 1.4291
No log 6.1739 142 1.6154 0.5193 1.6154 1.2710
No log 6.2609 144 0.9343 0.625 0.9343 0.9666
No log 6.3478 146 0.7183 0.7050 0.7183 0.8475
No log 6.4348 148 0.9153 0.6119 0.9153 0.9567
No log 6.5217 150 0.9873 0.6107 0.9873 0.9936
No log 6.6087 152 0.9539 0.5909 0.9539 0.9767
No log 6.6957 154 0.8704 0.6619 0.8704 0.9330
No log 6.7826 156 0.9048 0.6569 0.9048 0.9512
No log 6.8696 158 0.9014 0.6154 0.9014 0.9494
No log 6.9565 160 0.8569 0.6519 0.8569 0.9257
No log 7.0435 162 0.8744 0.6423 0.8744 0.9351
No log 7.1304 164 0.9441 0.6165 0.9441 0.9716
No log 7.2174 166 0.9397 0.6165 0.9397 0.9694
No log 7.3043 168 0.9643 0.5323 0.9643 0.9820
No log 7.3913 170 1.1121 0.5385 1.1121 1.0545
No log 7.4783 172 1.1327 0.5344 1.1327 1.0643
No log 7.5652 174 0.9689 0.6099 0.9689 0.9843
No log 7.6522 176 0.8749 0.6525 0.8749 0.9354
No log 7.7391 178 0.9474 0.6294 0.9474 0.9733
No log 7.8261 180 1.1058 0.5578 1.1058 1.0516
No log 7.9130 182 1.4256 0.3889 1.4256 1.1940
No log 8.0 184 2.0375 0.2429 2.0375 1.4274
No log 8.0870 186 1.8348 0.2963 1.8348 1.3546
No log 8.1739 188 1.2346 0.5271 1.2346 1.1111
No log 8.2609 190 1.0544 0.5714 1.0544 1.0268
No log 8.3478 192 0.9895 0.5926 0.9895 0.9947
No log 8.4348 194 0.9836 0.6377 0.9836 0.9918
No log 8.5217 196 0.9746 0.5649 0.9746 0.9872
No log 8.6087 198 0.9681 0.5649 0.9681 0.9839
No log 8.6957 200 1.0161 0.6429 1.0161 1.0080
No log 8.7826 202 1.1600 0.5672 1.1600 1.0770
No log 8.8696 204 1.1827 0.5414 1.1827 1.0875
No log 8.9565 206 1.0554 0.5197 1.0554 1.0273
No log 9.0435 208 1.0017 0.5625 1.0017 1.0008
No log 9.1304 210 0.9975 0.5625 0.9975 0.9987
No log 9.2174 212 1.0159 0.5714 1.0159 1.0079
No log 9.3043 214 0.9891 0.5693 0.9891 0.9945
No log 9.3913 216 0.9031 0.6294 0.9031 0.9503
No log 9.4783 218 0.8197 0.7092 0.8197 0.9054
No log 9.5652 220 0.8192 0.7413 0.8192 0.9051
No log 9.6522 222 0.8336 0.6853 0.8336 0.9130
No log 9.7391 224 0.8778 0.6528 0.8778 0.9369
No log 9.8261 226 0.9382 0.5833 0.9382 0.9686
No log 9.9130 228 0.9293 0.6197 0.9293 0.9640
No log 10.0 230 0.9180 0.6569 0.9180 0.9581
No log 10.0870 232 0.9080 0.6618 0.9080 0.9529
No log 10.1739 234 0.8941 0.6519 0.8941 0.9455
No log 10.2609 236 0.9193 0.6471 0.9193 0.9588
No log 10.3478 238 0.9469 0.6370 0.9469 0.9731
No log 10.4348 240 1.0144 0.5758 1.0144 1.0072
No log 10.5217 242 0.9638 0.6119 0.9638 0.9817
No log 10.6087 244 0.8408 0.6809 0.8408 0.9169
No log 10.6957 246 0.7935 0.7034 0.7935 0.8908
No log 10.7826 248 0.8039 0.6849 0.8039 0.8966
No log 10.8696 250 0.8587 0.6623 0.8587 0.9267
No log 10.9565 252 0.8617 0.6623 0.8617 0.9283
No log 11.0435 254 0.8394 0.6944 0.8394 0.9162
No log 11.1304 256 0.8746 0.7092 0.8746 0.9352
No log 11.2174 258 0.8890 0.7 0.8890 0.9428
No log 11.3043 260 0.8865 0.6761 0.8865 0.9415
No log 11.3913 262 0.9436 0.6187 0.9436 0.9714
No log 11.4783 264 0.9589 0.5942 0.9589 0.9792
No log 11.5652 266 0.9325 0.6316 0.9325 0.9656
No log 11.6522 268 0.9066 0.6618 0.9066 0.9521
No log 11.7391 270 0.8843 0.6715 0.8843 0.9404
No log 11.8261 272 0.8482 0.6812 0.8482 0.9210
No log 11.9130 274 0.8409 0.6812 0.8409 0.9170
No log 12.0 276 0.9243 0.6165 0.9243 0.9614
No log 12.0870 278 0.9715 0.6015 0.9715 0.9856
No log 12.1739 280 0.9139 0.6418 0.9139 0.9560
No log 12.2609 282 0.8167 0.7194 0.8167 0.9037
No log 12.3478 284 0.8005 0.7234 0.8005 0.8947
No log 12.4348 286 0.8313 0.7042 0.8313 0.9118
No log 12.5217 288 0.8357 0.6849 0.8357 0.9141
No log 12.6087 290 0.8818 0.6275 0.8818 0.9390
No log 12.6957 292 0.8860 0.6133 0.8860 0.9413
No log 12.7826 294 0.8119 0.7143 0.8119 0.9010
No log 12.8696 296 0.8093 0.7448 0.8093 0.8996
No log 12.9565 298 0.8377 0.7143 0.8377 0.9153
No log 13.0435 300 0.8772 0.6912 0.8772 0.9366
No log 13.1304 302 0.9210 0.5909 0.9210 0.9597
No log 13.2174 304 0.9069 0.6418 0.9069 0.9523
No log 13.3043 306 0.8712 0.7050 0.8712 0.9334
No log 13.3913 308 0.8834 0.6761 0.8834 0.9399
No log 13.4783 310 0.8619 0.6761 0.8619 0.9284
No log 13.5652 312 0.8309 0.7050 0.8309 0.9115
No log 13.6522 314 0.8812 0.6667 0.8812 0.9387
No log 13.7391 316 0.9264 0.6184 0.9264 0.9625
No log 13.8261 318 0.9201 0.6483 0.9201 0.9592
No log 13.9130 320 0.9213 0.6377 0.9213 0.9598
No log 14.0 322 0.9027 0.6377 0.9027 0.9501
No log 14.0870 324 0.8969 0.6377 0.8969 0.9470
No log 14.1739 326 0.9885 0.6099 0.9885 0.9942
No log 14.2609 328 1.2038 0.5714 1.2038 1.0972
No log 14.3478 330 1.1787 0.5714 1.1787 1.0857
No log 14.4348 332 1.0174 0.5828 1.0174 1.0087
No log 14.5217 334 0.8898 0.6525 0.8898 0.9433
No log 14.6087 336 0.8528 0.7101 0.8528 0.9235
No log 14.6957 338 0.8707 0.6957 0.8707 0.9331
No log 14.7826 340 0.8866 0.6912 0.8866 0.9416
No log 14.8696 342 0.9136 0.6667 0.9136 0.9558
No log 14.9565 344 0.9350 0.6569 0.9350 0.9670
No log 15.0435 346 0.9562 0.5758 0.9562 0.9778
No log 15.1304 348 0.9563 0.6154 0.9563 0.9779
No log 15.2174 350 0.9472 0.6154 0.9472 0.9732
No log 15.3043 352 0.9377 0.6260 0.9377 0.9683
No log 15.3913 354 0.9141 0.6269 0.9141 0.9561
No log 15.4783 356 0.9218 0.6222 0.9218 0.9601
No log 15.5652 358 0.9511 0.6176 0.9511 0.9752
No log 15.6522 360 0.9854 0.6294 0.9854 0.9927
No log 15.7391 362 1.0451 0.6323 1.0451 1.0223
No log 15.8261 364 1.0339 0.6323 1.0339 1.0168
No log 15.9130 366 0.9828 0.6104 0.9828 0.9914
No log 16.0 368 1.0053 0.6323 1.0053 1.0026
No log 16.0870 370 1.0107 0.5986 1.0107 1.0054
No log 16.1739 372 0.9142 0.6099 0.9142 0.9561
No log 16.2609 374 0.8494 0.6519 0.8494 0.9216
No log 16.3478 376 0.8456 0.6667 0.8456 0.9196
No log 16.4348 378 0.8768 0.6269 0.8768 0.9364
No log 16.5217 380 0.9361 0.6029 0.9361 0.9675
No log 16.6087 382 1.0166 0.5797 1.0166 1.0083
No log 16.6957 384 0.9900 0.5693 0.9900 0.9950
No log 16.7826 386 0.8943 0.6364 0.8943 0.9457
No log 16.8696 388 0.8525 0.6861 0.8525 0.9233
No log 16.9565 390 0.8555 0.6618 0.8555 0.9249
No log 17.0435 392 0.8608 0.6119 0.8608 0.9278
No log 17.1304 394 0.8809 0.5970 0.8809 0.9386
No log 17.2174 396 0.9873 0.5970 0.9873 0.9936
No log 17.3043 398 1.1589 0.6027 1.1589 1.0765
No log 17.3913 400 1.1617 0.5949 1.1617 1.0778
No log 17.4783 402 1.0279 0.5921 1.0279 1.0139
No log 17.5652 404 0.8699 0.6522 0.8699 0.9327
No log 17.6522 406 0.7673 0.6861 0.7673 0.8759
No log 17.7391 408 0.7499 0.7194 0.7499 0.8660
No log 17.8261 410 0.7567 0.6957 0.7567 0.8699
No log 17.9130 412 0.8043 0.6519 0.8043 0.8968
No log 18.0 414 0.8655 0.6667 0.8655 0.9303
No log 18.0870 416 0.9575 0.6429 0.9575 0.9785
No log 18.1739 418 0.9576 0.6667 0.9576 0.9785
No log 18.2609 420 0.8833 0.6277 0.8833 0.9398
No log 18.3478 422 0.8654 0.6370 0.8654 0.9303
No log 18.4348 424 0.8487 0.6618 0.8487 0.9212
No log 18.5217 426 0.8284 0.6861 0.8284 0.9102
No log 18.6087 428 0.8201 0.7050 0.8201 0.9056
No log 18.6957 430 0.8178 0.6522 0.8178 0.9043
No log 18.7826 432 0.8508 0.6980 0.8508 0.9224
No log 18.8696 434 0.8386 0.6928 0.8386 0.9158
No log 18.9565 436 0.8055 0.6711 0.8055 0.8975
No log 19.0435 438 0.8091 0.7067 0.8091 0.8995
No log 19.1304 440 0.8629 0.6928 0.8629 0.9289
No log 19.2174 442 0.8858 0.6933 0.8858 0.9412
No log 19.3043 444 0.8670 0.6711 0.8670 0.9311
No log 19.3913 446 0.8508 0.6575 0.8508 0.9224
No log 19.4783 448 0.8241 0.6154 0.8241 0.9078
No log 19.5652 450 0.8073 0.6892 0.8073 0.8985
No log 19.6522 452 0.7719 0.7222 0.7719 0.8786
No log 19.7391 454 0.7572 0.7222 0.7572 0.8702
No log 19.8261 456 0.7576 0.7222 0.7576 0.8704
No log 19.9130 458 0.7694 0.7133 0.7694 0.8771
No log 20.0 460 0.7978 0.7143 0.7978 0.8932
No log 20.0870 462 0.8447 0.6370 0.8447 0.9191
No log 20.1739 464 0.8614 0.6316 0.8614 0.9281
No log 20.2609 466 0.8503 0.6866 0.8503 0.9221
No log 20.3478 468 0.8366 0.7015 0.8366 0.9147
No log 20.4348 470 0.8186 0.6765 0.8186 0.9048
No log 20.5217 472 0.7829 0.7299 0.7829 0.8848
No log 20.6087 474 0.7718 0.7429 0.7718 0.8785
No log 20.6957 476 0.7844 0.6950 0.7844 0.8857
No log 20.7826 478 0.8550 0.6174 0.8550 0.9247
No log 20.8696 480 0.9567 0.6708 0.9567 0.9781
No log 20.9565 482 0.9710 0.6329 0.9710 0.9854
No log 21.0435 484 0.9619 0.6174 0.9619 0.9808
No log 21.1304 486 0.9518 0.6131 0.9518 0.9756
No log 21.2174 488 0.9437 0.6165 0.9437 0.9714
No log 21.3043 490 0.9472 0.6165 0.9472 0.9733
No log 21.3913 492 0.9551 0.6260 0.9551 0.9773
No log 21.4783 494 0.9581 0.6260 0.9581 0.9788
No log 21.5652 496 0.9537 0.6418 0.9537 0.9766
No log 21.6522 498 0.9287 0.6618 0.9287 0.9637
0.3585 21.7391 500 0.9021 0.6765 0.9021 0.9498
0.3585 21.8261 502 0.8828 0.6618 0.8828 0.9396
0.3585 21.9130 504 0.8642 0.7234 0.8642 0.9296
0.3585 22.0 506 0.8503 0.7042 0.8503 0.9221
0.3585 22.0870 508 0.8291 0.7042 0.8291 0.9105
0.3585 22.1739 510 0.8020 0.6857 0.8020 0.8955
0.3585 22.2609 512 0.8101 0.6906 0.8101 0.9000
0.3585 22.3478 514 0.8738 0.6763 0.8738 0.9348
0.3585 22.4348 516 0.8720 0.6763 0.8720 0.9338
0.3585 22.5217 518 0.8325 0.6812 0.8325 0.9124
0.3585 22.6087 520 0.8124 0.6565 0.8124 0.9013
0.3585 22.6957 522 0.8228 0.6906 0.8228 0.9071
0.3585 22.7826 524 0.8327 0.7183 0.8327 0.9125
0.3585 22.8696 526 0.8404 0.7007 0.8404 0.9167
0.3585 22.9565 528 0.8658 0.6519 0.8658 0.9305
0.3585 23.0435 530 0.9097 0.5899 0.9097 0.9538
0.3585 23.1304 532 0.9383 0.5753 0.9383 0.9686
0.3585 23.2174 534 0.8987 0.6099 0.8987 0.9480
0.3585 23.3043 536 0.8523 0.6619 0.8523 0.9232

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k5_task1_organization

Finetuned
(4023)
this model