ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k11_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7917
  • Qwk: 0.6857
  • Mse: 0.7917
  • Rmse: 0.8898

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0392 2 7.1782 -0.0167 7.1782 2.6792
No log 0.0784 4 5.1691 0.0 5.1691 2.2736
No log 0.1176 6 4.1512 -0.0580 4.1512 2.0374
No log 0.1569 8 2.4739 0.1286 2.4739 1.5729
No log 0.1961 10 1.8534 0.1724 1.8534 1.3614
No log 0.2353 12 1.8446 0.1607 1.8446 1.3582
No log 0.2745 14 2.2032 0.0640 2.2032 1.4843
No log 0.3137 16 2.1366 0.0833 2.1366 1.4617
No log 0.3529 18 2.1681 0.1626 2.1681 1.4725
No log 0.3922 20 2.1985 0.1406 2.1985 1.4827
No log 0.4314 22 2.1276 0.2031 2.1276 1.4586
No log 0.4706 24 2.0415 0.2424 2.0415 1.4288
No log 0.5098 26 1.9496 0.3140 1.9496 1.3963
No log 0.5490 28 1.7358 0.2407 1.7358 1.3175
No log 0.5882 30 1.5554 0.1869 1.5554 1.2472
No log 0.6275 32 1.6556 0.2261 1.6556 1.2867
No log 0.6667 34 1.9373 0.2656 1.9373 1.3919
No log 0.7059 36 1.8701 0.2687 1.8701 1.3675
No log 0.7451 38 1.5227 0.3465 1.5227 1.2340
No log 0.7843 40 1.2858 0.4839 1.2858 1.1339
No log 0.8235 42 1.2447 0.5410 1.2447 1.1157
No log 0.8627 44 1.2360 0.4590 1.2360 1.1117
No log 0.9020 46 1.3770 0.3438 1.3770 1.1735
No log 0.9412 48 1.9377 0.3273 1.9377 1.3920
No log 0.9804 50 2.1695 0.3459 2.1695 1.4729
No log 1.0196 52 2.1500 0.3511 2.1500 1.4663
No log 1.0588 54 1.7957 0.3855 1.7957 1.3400
No log 1.0980 56 1.5940 0.3862 1.5940 1.2626
No log 1.1373 58 1.4410 0.4706 1.4410 1.2004
No log 1.1765 60 1.8067 0.3586 1.8067 1.3442
No log 1.2157 62 2.0149 0.2177 2.0149 1.4195
No log 1.2549 64 1.9243 0.2897 1.9243 1.3872
No log 1.2941 66 1.4111 0.5294 1.4111 1.1879
No log 1.3333 68 1.2605 0.5075 1.2605 1.1227
No log 1.3725 70 1.1855 0.4844 1.1855 1.0888
No log 1.4118 72 1.2190 0.5038 1.2190 1.1041
No log 1.4510 74 1.0329 0.5079 1.0329 1.0163
No log 1.4902 76 1.0197 0.5538 1.0197 1.0098
No log 1.5294 78 1.0213 0.5954 1.0213 1.0106
No log 1.5686 80 1.0808 0.6119 1.0808 1.0396
No log 1.6078 82 1.1093 0.6015 1.1093 1.0533
No log 1.6471 84 1.1167 0.5606 1.1167 1.0568
No log 1.6863 86 1.1311 0.5362 1.1311 1.0635
No log 1.7255 88 1.1852 0.5769 1.1852 1.0887
No log 1.7647 90 1.1023 0.5630 1.1023 1.0499
No log 1.8039 92 1.7678 0.2519 1.7678 1.3296
No log 1.8431 94 2.8107 -0.0916 2.8107 1.6765
No log 1.8824 96 3.4161 -0.1727 3.4161 1.8483
No log 1.9216 98 2.0707 0.1575 2.0707 1.4390
No log 1.9608 100 0.8760 0.6377 0.8760 0.9359
No log 2.0 102 1.1913 0.5857 1.1913 1.0915
No log 2.0392 104 1.8752 0.3590 1.8752 1.3694
No log 2.0784 106 1.7041 0.3871 1.7041 1.3054
No log 2.1176 108 0.9763 0.6099 0.9763 0.9881
No log 2.1569 110 0.8975 0.6815 0.8975 0.9474
No log 2.1961 112 1.0856 0.5692 1.0856 1.0419
No log 2.2353 114 1.0253 0.5954 1.0253 1.0126
No log 2.2745 116 1.1049 0.5865 1.1049 1.0512
No log 2.3137 118 0.9565 0.6222 0.9565 0.9780
No log 2.3529 120 0.8230 0.7273 0.8230 0.9072
No log 2.3922 122 0.7788 0.7237 0.7788 0.8825
No log 2.4314 124 0.7793 0.6939 0.7793 0.8828
No log 2.4706 126 0.7809 0.6765 0.7809 0.8837
No log 2.5098 128 0.7972 0.6812 0.7972 0.8928
No log 2.5490 130 0.8233 0.6715 0.8233 0.9074
No log 2.5882 132 0.8861 0.5926 0.8861 0.9413
No log 2.6275 134 0.9127 0.5970 0.9127 0.9554
No log 2.6667 136 0.9450 0.5397 0.9450 0.9721
No log 2.7059 138 1.0962 0.5691 1.0962 1.0470
No log 2.7451 140 1.1291 0.5289 1.1291 1.0626
No log 2.7843 142 1.1377 0.4228 1.1377 1.0666
No log 2.8235 144 1.1790 0.4812 1.1790 1.0858
No log 2.8627 146 1.0789 0.5857 1.0789 1.0387
No log 2.9020 148 0.9969 0.6620 0.9969 0.9985
No log 2.9412 150 1.3038 0.5224 1.3038 1.1418
No log 2.9804 152 1.3741 0.4812 1.3741 1.1722
No log 3.0196 154 1.0429 0.6222 1.0429 1.0212
No log 3.0588 156 0.8989 0.6423 0.8989 0.9481
No log 3.0980 158 0.9039 0.6569 0.9039 0.9507
No log 3.1373 160 0.9325 0.6176 0.9325 0.9657
No log 3.1765 162 0.9084 0.6176 0.9084 0.9531
No log 3.2157 164 0.8983 0.5926 0.8983 0.9478
No log 3.2549 166 0.8832 0.6074 0.8832 0.9398
No log 3.2941 168 0.8754 0.6715 0.8754 0.9356
No log 3.3333 170 0.8934 0.6232 0.8934 0.9452
No log 3.3725 172 0.8340 0.7 0.8340 0.9133
No log 3.4118 174 0.7807 0.7172 0.7807 0.8835
No log 3.4510 176 0.8629 0.6906 0.8629 0.9289
No log 3.4902 178 1.0104 0.6618 1.0104 1.0052
No log 3.5294 180 1.0748 0.6316 1.0748 1.0367
No log 3.5686 182 1.1415 0.6107 1.1415 1.0684
No log 3.6078 184 1.0634 0.6567 1.0634 1.0312
No log 3.6471 186 0.9877 0.6617 0.9877 0.9938
No log 3.6863 188 0.9425 0.6047 0.9425 0.9708
No log 3.7255 190 0.9274 0.5581 0.9274 0.9630
No log 3.7647 192 0.9008 0.6165 0.9008 0.9491
No log 3.8039 194 0.8503 0.6423 0.8503 0.9221
No log 3.8431 196 0.7972 0.7083 0.7972 0.8928
No log 3.8824 198 0.7649 0.7172 0.7649 0.8746
No log 3.9216 200 0.7621 0.7333 0.7621 0.8730
No log 3.9608 202 0.7512 0.7383 0.7512 0.8667
No log 4.0 204 0.8115 0.6761 0.8115 0.9009
No log 4.0392 206 0.8588 0.6471 0.8588 0.9267
No log 4.0784 208 0.8957 0.6522 0.8957 0.9464
No log 4.1176 210 0.9823 0.5926 0.9823 0.9911
No log 4.1569 212 0.9946 0.5714 0.9946 0.9973
No log 4.1961 214 1.0250 0.5714 1.0250 1.0124
No log 4.2353 216 0.9861 0.5538 0.9861 0.9930
No log 4.2745 218 0.9968 0.6316 0.9968 0.9984
No log 4.3137 220 1.0043 0.6316 1.0043 1.0022
No log 4.3529 222 0.9641 0.6061 0.9641 0.9819
No log 4.3922 224 0.9178 0.5846 0.9178 0.9580
No log 4.4314 226 0.8756 0.6475 0.8756 0.9357
No log 4.4706 228 0.8524 0.6809 0.8524 0.9232
No log 4.5098 230 0.9128 0.6277 0.9128 0.9554
No log 4.5490 232 0.9419 0.6232 0.9419 0.9705
No log 4.5882 234 0.9130 0.6277 0.9130 0.9555
No log 4.6275 236 0.9071 0.6269 0.9071 0.9524
No log 4.6667 238 0.9396 0.6131 0.9396 0.9693
No log 4.7059 240 0.9246 0.6294 0.9246 0.9615
No log 4.7451 242 0.8222 0.7172 0.8222 0.9068
No log 4.7843 244 0.9092 0.6573 0.9092 0.9535
No log 4.8235 246 0.9489 0.6377 0.9489 0.9741
No log 4.8627 248 0.9240 0.6370 0.9240 0.9612
No log 4.9020 250 1.0074 0.6331 1.0074 1.0037
No log 4.9412 252 1.1457 0.5306 1.1457 1.0704
No log 4.9804 254 1.1316 0.5658 1.1316 1.0638
No log 5.0196 256 1.0444 0.5634 1.0444 1.0220
No log 5.0588 258 0.9850 0.512 0.9850 0.9924
No log 5.0980 260 0.9323 0.6212 0.9323 0.9656
No log 5.1373 262 0.9045 0.6765 0.9045 0.9511
No log 5.1765 264 0.8365 0.6763 0.8365 0.9146
No log 5.2157 266 0.8368 0.6713 0.8368 0.9148
No log 5.2549 268 0.8899 0.6497 0.8899 0.9434
No log 5.2941 270 0.8736 0.6447 0.8736 0.9347
No log 5.3333 272 0.8279 0.6761 0.8279 0.9099
No log 5.3725 274 0.8192 0.6715 0.8192 0.9051
No log 5.4118 276 0.8298 0.6519 0.8298 0.9110
No log 5.4510 278 0.8128 0.6912 0.8128 0.9016
No log 5.4902 280 0.8098 0.6950 0.8098 0.8999
No log 5.5294 282 0.8865 0.6197 0.8865 0.9416
No log 5.5686 284 0.8317 0.6667 0.8317 0.9120
No log 5.6078 286 0.7545 0.7123 0.7545 0.8686
No log 5.6471 288 0.7744 0.7042 0.7744 0.8800
No log 5.6863 290 0.8289 0.7042 0.8289 0.9104
No log 5.7255 292 0.8815 0.6857 0.8815 0.9389
No log 5.7647 294 0.9484 0.6119 0.9484 0.9739
No log 5.8039 296 0.9549 0.5865 0.9549 0.9772
No log 5.8431 298 0.9289 0.5970 0.9289 0.9638
No log 5.8824 300 0.8705 0.5926 0.8705 0.9330
No log 5.9216 302 0.7947 0.6667 0.7947 0.8915
No log 5.9608 304 0.8332 0.7162 0.8332 0.9128
No log 6.0 306 0.9310 0.6438 0.9310 0.9649
No log 6.0392 308 1.0507 0.6187 1.0507 1.0250
No log 6.0784 310 1.0085 0.6528 1.0085 1.0043
No log 6.1176 312 0.7897 0.7397 0.7897 0.8887
No log 6.1569 314 0.7442 0.7133 0.7442 0.8626
No log 6.1961 316 0.9297 0.6 0.9297 0.9642
No log 6.2353 318 0.9226 0.5778 0.9226 0.9605
No log 6.2745 320 0.7970 0.6950 0.7970 0.8928
No log 6.3137 322 0.8181 0.7143 0.8181 0.9045
No log 6.3529 324 0.8071 0.7347 0.8071 0.8984
No log 6.3922 326 0.7857 0.75 0.7857 0.8864
No log 6.4314 328 0.8047 0.7237 0.8047 0.8970
No log 6.4706 330 0.7894 0.7467 0.7894 0.8885
No log 6.5098 332 0.7992 0.6887 0.7992 0.8940
No log 6.5490 334 0.8165 0.6803 0.8165 0.9036
No log 6.5882 336 0.8518 0.6714 0.8518 0.9229
No log 6.6275 338 0.8914 0.6569 0.8914 0.9441
No log 6.6667 340 0.9228 0.6119 0.9228 0.9606
No log 6.7059 342 0.9146 0.6277 0.9146 0.9563
No log 6.7451 344 0.8456 0.6897 0.8456 0.9196
No log 6.7843 346 0.8381 0.6846 0.8381 0.9155
No log 6.8235 348 0.8291 0.6711 0.8291 0.9106
No log 6.8627 350 0.8298 0.72 0.8298 0.9109
No log 6.9020 352 0.9582 0.6713 0.9582 0.9789
No log 6.9412 354 0.9528 0.6569 0.9528 0.9761
No log 6.9804 356 0.8821 0.6269 0.8821 0.9392
No log 7.0196 358 0.9396 0.6338 0.9396 0.9693
No log 7.0588 360 0.9785 0.5714 0.9785 0.9892
No log 7.0980 362 0.8457 0.6620 0.8457 0.9196
No log 7.1373 364 0.8037 0.6812 0.8037 0.8965
No log 7.1765 366 0.9130 0.6571 0.9130 0.9555
No log 7.2157 368 1.0068 0.6099 1.0068 1.0034
No log 7.2549 370 0.9148 0.6763 0.9148 0.9564
No log 7.2941 372 0.7621 0.75 0.7621 0.8730
No log 7.3333 374 0.7503 0.7248 0.7503 0.8662
No log 7.3725 376 0.7357 0.72 0.7357 0.8577
No log 7.4118 378 0.7088 0.7568 0.7088 0.8419
No log 7.4510 380 0.8207 0.7172 0.8207 0.9059
No log 7.4902 382 0.8776 0.6901 0.8776 0.9368
No log 7.5294 384 0.8945 0.6763 0.8945 0.9458
No log 7.5686 386 0.7576 0.7222 0.7576 0.8704
No log 7.6078 388 0.7452 0.7 0.7452 0.8633
No log 7.6471 390 0.7993 0.7123 0.7993 0.8940
No log 7.6863 392 0.7970 0.6715 0.7970 0.8928
No log 7.7255 394 0.7895 0.6714 0.7895 0.8885
No log 7.7647 396 0.7954 0.6944 0.7954 0.8918
No log 7.8039 398 0.7946 0.6950 0.7946 0.8914
No log 7.8431 400 0.8101 0.6809 0.8101 0.9000
No log 7.8824 402 0.8026 0.6809 0.8026 0.8959
No log 7.9216 404 0.8070 0.6901 0.8070 0.8983
No log 7.9608 406 0.8007 0.6809 0.8007 0.8948
No log 8.0 408 0.8332 0.6806 0.8332 0.9128
No log 8.0392 410 0.8401 0.6759 0.8401 0.9166
No log 8.0784 412 0.7928 0.6806 0.7928 0.8904
No log 8.1176 414 0.7651 0.6812 0.7651 0.8747
No log 8.1569 416 0.8211 0.6812 0.8211 0.9062
No log 8.1961 418 0.8925 0.6812 0.8925 0.9447
No log 8.2353 420 0.8935 0.6763 0.8935 0.9453
No log 8.2745 422 0.8237 0.6950 0.8237 0.9076
No log 8.3137 424 0.7780 0.7347 0.7780 0.8820
No log 8.3529 426 0.7845 0.7260 0.7845 0.8857
No log 8.3922 428 0.8093 0.7397 0.8093 0.8996
No log 8.4314 430 0.8428 0.7092 0.8428 0.9181
No log 8.4706 432 0.8588 0.6861 0.8588 0.9267
No log 8.5098 434 0.8601 0.6667 0.8601 0.9274
No log 8.5490 436 0.8628 0.6667 0.8628 0.9289
No log 8.5882 438 0.8308 0.6715 0.8308 0.9115
No log 8.6275 440 0.7983 0.7050 0.7983 0.8935
No log 8.6667 442 0.7533 0.7483 0.7533 0.8679
No log 8.7059 444 0.7471 0.7651 0.7471 0.8643
No log 8.7451 446 0.7670 0.7347 0.7670 0.8758
No log 8.7843 448 0.7977 0.7260 0.7977 0.8932
No log 8.8235 450 0.7830 0.7133 0.7830 0.8849
No log 8.8627 452 0.7391 0.7651 0.7391 0.8597
No log 8.9020 454 0.7704 0.7483 0.7704 0.8777
No log 8.9412 456 0.7880 0.7172 0.7880 0.8877
No log 8.9804 458 0.8067 0.7222 0.8067 0.8981
No log 9.0196 460 0.8155 0.6857 0.8155 0.9031
No log 9.0588 462 0.8811 0.6861 0.8811 0.9387
No log 9.0980 464 0.8978 0.6618 0.8978 0.9475
No log 9.1373 466 0.8766 0.6667 0.8766 0.9363
No log 9.1765 468 0.8750 0.6569 0.8750 0.9354
No log 9.2157 470 0.8902 0.6812 0.8902 0.9435
No log 9.2549 472 0.9084 0.6667 0.9084 0.9531
No log 9.2941 474 0.9169 0.6187 0.9169 0.9575
No log 9.3333 476 0.8721 0.6906 0.8721 0.9339
No log 9.3725 478 0.8336 0.6857 0.8336 0.9130
No log 9.4118 480 0.8150 0.7172 0.8150 0.9028
No log 9.4510 482 0.8118 0.6857 0.8118 0.9010
No log 9.4902 484 0.8563 0.6906 0.8563 0.9254
No log 9.5294 486 0.9065 0.6763 0.9065 0.9521
No log 9.5686 488 0.9388 0.6187 0.9388 0.9689
No log 9.6078 490 0.9461 0.6522 0.9461 0.9727
No log 9.6471 492 0.8685 0.7 0.8685 0.9319
No log 9.6863 494 0.8205 0.7034 0.8205 0.9058
No log 9.7255 496 0.8115 0.6842 0.8115 0.9008
No log 9.7647 498 0.7997 0.7342 0.7997 0.8943
0.4048 9.8039 500 0.7948 0.7114 0.7948 0.8915
0.4048 9.8431 502 0.8147 0.7285 0.8147 0.9026
0.4048 9.8824 504 0.8288 0.7273 0.8288 0.9104
0.4048 9.9216 506 0.7974 0.7 0.7974 0.8929
0.4048 9.9608 508 0.7885 0.6857 0.7885 0.8880
0.4048 10.0 510 0.8318 0.6429 0.8318 0.9120
0.4048 10.0392 512 0.8362 0.6761 0.8362 0.9145
0.4048 10.0784 514 0.7292 0.7133 0.7292 0.8539
0.4048 10.1176 516 0.6814 0.7448 0.6814 0.8255
0.4048 10.1569 518 0.6758 0.7619 0.6758 0.8221
0.4048 10.1961 520 0.6992 0.7619 0.6992 0.8362
0.4048 10.2353 522 0.7609 0.7273 0.7609 0.8723
0.4048 10.2745 524 0.7549 0.7273 0.7549 0.8688
0.4048 10.3137 526 0.7669 0.7183 0.7669 0.8758
0.4048 10.3529 528 0.8056 0.7050 0.8056 0.8976
0.4048 10.3922 530 0.8057 0.6957 0.8057 0.8976
0.4048 10.4314 532 0.8117 0.7050 0.8117 0.9009
0.4048 10.4706 534 0.8038 0.6857 0.8038 0.8965
0.4048 10.5098 536 0.7917 0.6857 0.7917 0.8898

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k11_task1_organization

Finetuned
(4019)
this model