ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k13_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7755
  • Qwk: 0.0863
  • Mse: 0.7755
  • Rmse: 0.8806

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0556 2 3.4981 0.0183 3.4981 1.8703
No log 0.1111 4 1.9415 0.0304 1.9415 1.3934
No log 0.1667 6 2.2890 -0.0129 2.2890 1.5129
No log 0.2222 8 1.0720 -0.1596 1.0720 1.0354
No log 0.2778 10 0.7585 0.1259 0.7585 0.8709
No log 0.3333 12 0.6890 0.0555 0.6890 0.8301
No log 0.3889 14 0.7427 0.1021 0.7427 0.8618
No log 0.4444 16 1.0700 -0.0468 1.0700 1.0344
No log 0.5 18 1.4275 0.0 1.4275 1.1948
No log 0.5556 20 1.3197 -0.0234 1.3197 1.1488
No log 0.6111 22 0.9274 -0.0638 0.9274 0.9630
No log 0.6667 24 0.7341 0.0334 0.7341 0.8568
No log 0.7222 26 0.6966 0.0964 0.6966 0.8346
No log 0.7778 28 0.7173 0.0374 0.7173 0.8469
No log 0.8333 30 0.8026 -0.0287 0.8026 0.8959
No log 0.8889 32 1.1066 -0.0234 1.1066 1.0519
No log 0.9444 34 1.2878 0.0 1.2878 1.1348
No log 1.0 36 1.1675 0.0 1.1675 1.0805
No log 1.0556 38 0.9496 -0.0385 0.9496 0.9745
No log 1.1111 40 0.8817 -0.0474 0.8817 0.9390
No log 1.1667 42 0.8042 0.0296 0.8042 0.8967
No log 1.2222 44 0.7306 0.0 0.7306 0.8548
No log 1.2778 46 0.7380 0.0 0.7380 0.8590
No log 1.3333 48 0.8915 -0.0033 0.8915 0.9442
No log 1.3889 50 1.6730 -0.0234 1.6730 1.2934
No log 1.4444 52 2.0502 -0.0014 2.0502 1.4319
No log 1.5 54 1.6190 -0.0234 1.6190 1.2724
No log 1.5556 56 1.1034 -0.0149 1.1034 1.0504
No log 1.6111 58 0.9689 0.0329 0.9689 0.9843
No log 1.6667 60 0.8987 0.0134 0.8987 0.9480
No log 1.7222 62 0.8577 0.0676 0.8577 0.9261
No log 1.7778 64 0.9090 0.0486 0.9090 0.9534
No log 1.8333 66 0.7975 0.0549 0.7975 0.8930
No log 1.8889 68 0.7076 -0.0035 0.7076 0.8412
No log 1.9444 70 0.7051 -0.0035 0.7051 0.8397
No log 2.0 72 0.7185 0.0 0.7185 0.8477
No log 2.0556 74 0.7353 -0.0033 0.7353 0.8575
No log 2.1111 76 0.7784 -0.0125 0.7784 0.8823
No log 2.1667 78 0.7859 -0.0204 0.7859 0.8865
No log 2.2222 80 0.9024 -0.0809 0.9024 0.9500
No log 2.2778 82 1.6677 -0.0235 1.6677 1.2914
No log 2.3333 84 2.2872 -0.0481 2.2872 1.5123
No log 2.3889 86 1.9063 -0.0014 1.9063 1.3807
No log 2.4444 88 1.2666 0.0196 1.2666 1.1254
No log 2.5 90 0.7963 -0.0675 0.7963 0.8924
No log 2.5556 92 0.7599 -0.1001 0.7599 0.8717
No log 2.6111 94 0.8193 -0.0939 0.8193 0.9051
No log 2.6667 96 0.7737 0.0099 0.7737 0.8796
No log 2.7222 98 0.7679 -0.0252 0.7679 0.8763
No log 2.7778 100 0.8099 -0.1628 0.8099 0.9000
No log 2.8333 102 0.7784 0.0978 0.7784 0.8823
No log 2.8889 104 0.7997 0.1030 0.7997 0.8943
No log 2.9444 106 0.8532 0.1188 0.8532 0.9237
No log 3.0 108 0.7784 0.0791 0.7784 0.8823
No log 3.0556 110 0.7725 0.1141 0.7725 0.8789
No log 3.1111 112 0.8250 0.0239 0.8250 0.9083
No log 3.1667 114 0.8327 0.0592 0.8327 0.9125
No log 3.2222 116 0.8372 -0.0573 0.8372 0.9150
No log 3.2778 118 0.7835 0.0976 0.7835 0.8851
No log 3.3333 120 0.7583 0.0585 0.7583 0.8708
No log 3.3889 122 0.7125 -0.0033 0.7125 0.8441
No log 3.4444 124 0.9164 -0.1261 0.9164 0.9573
No log 3.5 126 1.0421 -0.0892 1.0421 1.0208
No log 3.5556 128 0.7175 0.0909 0.7175 0.8471
No log 3.6111 130 0.7445 0.0064 0.7445 0.8628
No log 3.6667 132 0.8903 0.0789 0.8903 0.9435
No log 3.7222 134 0.8440 -0.0089 0.8440 0.9187
No log 3.7778 136 0.7692 0.0247 0.7692 0.8770
No log 3.8333 138 1.0331 0.0107 1.0331 1.0164
No log 3.8889 140 1.2342 0.0456 1.2342 1.1109
No log 3.9444 142 0.9054 0.0277 0.9054 0.9515
No log 4.0 144 0.8604 0.0279 0.8604 0.9276
No log 4.0556 146 1.0345 -0.0142 1.0345 1.0171
No log 4.1111 148 0.8901 -0.0015 0.8901 0.9434
No log 4.1667 150 0.8053 -0.0252 0.8053 0.8974
No log 4.2222 152 0.9325 0.0200 0.9325 0.9657
No log 4.2778 154 0.8415 -0.0355 0.8415 0.9174
No log 4.3333 156 0.7783 0.0723 0.7783 0.8822
No log 4.3889 158 0.7723 0.1541 0.7723 0.8788
No log 4.4444 160 0.8329 0.1286 0.8329 0.9126
No log 4.5 162 0.8097 0.1841 0.8097 0.8998
No log 4.5556 164 0.7762 -0.0634 0.7762 0.8810
No log 4.6111 166 0.7668 0.0187 0.7668 0.8757
No log 4.6667 168 0.7001 0.1081 0.7001 0.8367
No log 4.7222 170 0.6946 0.0814 0.6946 0.8335
No log 4.7778 172 0.7636 0.1286 0.7636 0.8739
No log 4.8333 174 0.7463 0.0600 0.7463 0.8639
No log 4.8889 176 0.7157 0.1081 0.7157 0.8460
No log 4.9444 178 0.7415 0.0732 0.7415 0.8611
No log 5.0 180 0.9104 0.0545 0.9104 0.9541
No log 5.0556 182 0.8706 0.0642 0.8706 0.9330
No log 5.1111 184 0.8508 0.1149 0.8508 0.9224
No log 5.1667 186 0.7998 0.1079 0.7998 0.8943
No log 5.2222 188 0.8390 0.0699 0.8390 0.9160
No log 5.2778 190 0.7898 0.1415 0.7898 0.8887
No log 5.3333 192 1.0441 0.0458 1.0441 1.0218
No log 5.3889 194 0.9615 0.0182 0.9615 0.9806
No log 5.4444 196 0.7847 0.1224 0.7847 0.8858
No log 5.5 198 0.9168 0.0443 0.9168 0.9575
No log 5.5556 200 0.8569 0.0431 0.8569 0.9257
No log 5.6111 202 0.7224 0.1501 0.7224 0.8500
No log 5.6667 204 0.6942 0.1878 0.6942 0.8332
No log 5.7222 206 0.7335 0.2181 0.7335 0.8564
No log 5.7778 208 0.8235 0.0248 0.8235 0.9075
No log 5.8333 210 1.0004 0.0764 1.0004 1.0002
No log 5.8889 212 0.9289 0.0239 0.9289 0.9638
No log 5.9444 214 0.8957 0.0654 0.8957 0.9464
No log 6.0 216 1.2158 0.0025 1.2158 1.1026
No log 6.0556 218 1.1945 0.0646 1.1945 1.0929
No log 6.1111 220 0.8172 0.0152 0.8172 0.9040
No log 6.1667 222 0.7356 0.1030 0.7356 0.8577
No log 6.2222 224 0.7104 0.2395 0.7104 0.8428
No log 6.2778 226 0.7684 0.1254 0.7684 0.8766
No log 6.3333 228 1.0208 0.0111 1.0208 1.0103
No log 6.3889 230 0.9696 -0.0532 0.9696 0.9847
No log 6.4444 232 0.7956 0.1097 0.7956 0.8920
No log 6.5 234 0.7813 0.0949 0.7813 0.8839
No log 6.5556 236 0.8104 0.0165 0.8104 0.9002
No log 6.6111 238 0.8248 0.0639 0.8248 0.9082
No log 6.6667 240 0.9630 -0.0143 0.9630 0.9813
No log 6.7222 242 0.8298 0.0068 0.8298 0.9109
No log 6.7778 244 0.7775 0.1986 0.7775 0.8818
No log 6.8333 246 0.7546 0.1561 0.7546 0.8687
No log 6.8889 248 0.7926 0.1943 0.7926 0.8903
No log 6.9444 250 0.8143 0.2208 0.8143 0.9024
No log 7.0 252 0.7205 0.2096 0.7205 0.8488
No log 7.0556 254 0.7081 0.1404 0.7081 0.8415
No log 7.1111 256 0.7661 0.2718 0.7661 0.8753
No log 7.1667 258 0.8280 0.0711 0.8280 0.9099
No log 7.2222 260 0.7456 0.2034 0.7456 0.8635
No log 7.2778 262 0.7241 0.1691 0.7241 0.8509
No log 7.3333 264 0.7158 0.0128 0.7158 0.8461
No log 7.3889 266 0.7899 -0.0633 0.7899 0.8888
No log 7.4444 268 0.7258 0.0615 0.7258 0.8520
No log 7.5 270 0.8615 0.0946 0.8615 0.9282
No log 7.5556 272 0.9844 -0.0218 0.9844 0.9922
No log 7.6111 274 0.7972 0.0793 0.7972 0.8929
No log 7.6667 276 0.7981 -0.0528 0.7981 0.8934
No log 7.7222 278 0.8875 0.0446 0.8875 0.9420
No log 7.7778 280 0.7526 -0.0293 0.7526 0.8675
No log 7.8333 282 0.7786 0.1742 0.7786 0.8824
No log 7.8889 284 0.7973 0.1235 0.7973 0.8929
No log 7.9444 286 0.7208 0.0357 0.7208 0.8490
No log 8.0 288 0.7195 0.0622 0.7195 0.8482
No log 8.0556 290 0.7018 0.1434 0.7018 0.8377
No log 8.1111 292 0.8579 0.0676 0.8579 0.9262
No log 8.1667 294 1.0432 0.0680 1.0432 1.0214
No log 8.2222 296 0.8984 0.0207 0.8984 0.9479
No log 8.2778 298 0.6919 0.0909 0.6919 0.8318
No log 8.3333 300 0.6717 0.1021 0.6717 0.8196
No log 8.3889 302 0.6835 0.0964 0.6835 0.8267
No log 8.4444 304 0.7851 0.1150 0.7851 0.8861
No log 8.5 306 0.8533 0.1024 0.8533 0.9238
No log 8.5556 308 0.8261 0.1841 0.8261 0.9089
No log 8.6111 310 0.7176 0.1705 0.7176 0.8471
No log 8.6667 312 0.7103 0.0798 0.7103 0.8428
No log 8.7222 314 0.7282 -0.0238 0.7282 0.8534
No log 8.7778 316 0.7322 -0.0217 0.7322 0.8557
No log 8.8333 318 0.7817 -0.0722 0.7817 0.8842
No log 8.8889 320 0.8000 -0.0685 0.8000 0.8944
No log 8.9444 322 0.7359 0.0187 0.7359 0.8579
No log 9.0 324 0.7982 0.0953 0.7982 0.8934
No log 9.0556 326 0.8580 0.1291 0.8580 0.9263
No log 9.1111 328 0.7893 0.0549 0.7893 0.8884
No log 9.1667 330 0.7688 0.1617 0.7688 0.8768
No log 9.2222 332 0.7879 0.0074 0.7879 0.8877
No log 9.2778 334 0.7845 0.0410 0.7845 0.8857
No log 9.3333 336 0.8346 0.0953 0.8346 0.9135
No log 9.3889 338 1.1539 0.0391 1.1539 1.0742
No log 9.4444 340 1.1265 0.0366 1.1265 1.0613
No log 9.5 342 0.8361 0.1196 0.8361 0.9144
No log 9.5556 344 0.7056 0.1498 0.7056 0.8400
No log 9.6111 346 0.7251 0.0155 0.7251 0.8515
No log 9.6667 348 0.7101 0.1878 0.7101 0.8426
No log 9.7222 350 0.7874 0.1660 0.7874 0.8874
No log 9.7778 352 0.9258 0.0609 0.9258 0.9622
No log 9.8333 354 0.8300 0.1660 0.8300 0.9110
No log 9.8889 356 0.7766 0.1617 0.7766 0.8813
No log 9.9444 358 0.8522 -0.0456 0.8522 0.9232
No log 10.0 360 0.8603 -0.0036 0.8603 0.9275
No log 10.0556 362 0.7965 0.0071 0.7965 0.8925
No log 10.1111 364 0.8144 0.0205 0.8144 0.9024
No log 10.1667 366 0.8006 0.0660 0.8006 0.8947
No log 10.2222 368 0.7653 0.1189 0.7653 0.8748
No log 10.2778 370 0.7544 -0.0320 0.7544 0.8685
No log 10.3333 372 0.7539 -0.0300 0.7539 0.8683
No log 10.3889 374 0.7264 0.1413 0.7264 0.8523
No log 10.4444 376 0.7132 0.0914 0.7132 0.8445
No log 10.5 378 0.7471 0.1097 0.7471 0.8643
No log 10.5556 380 0.8796 0.0642 0.8796 0.9378
No log 10.6111 382 0.8370 0.1291 0.8370 0.9149
No log 10.6667 384 0.7692 0.0308 0.7692 0.8770
No log 10.7222 386 0.8567 0.0696 0.8567 0.9256
No log 10.7778 388 0.8171 -0.0133 0.8171 0.9039
No log 10.8333 390 0.7539 0.0247 0.7539 0.8683
No log 10.8889 392 0.8673 0.0642 0.8673 0.9313
No log 10.9444 394 1.1805 0.0166 1.1805 1.0865
No log 11.0 396 1.2381 0.0204 1.2381 1.1127
No log 11.0556 398 0.9661 0.0786 0.9661 0.9829
No log 11.1111 400 0.7454 0.0814 0.7454 0.8634
No log 11.1667 402 0.7655 -0.0912 0.7655 0.8749
No log 11.2222 404 0.7487 -0.0446 0.7487 0.8653
No log 11.2778 406 0.7319 0.1318 0.7319 0.8555
No log 11.3333 408 0.7682 0.1716 0.7682 0.8765
No log 11.3889 410 0.7552 0.1395 0.7552 0.8690
No log 11.4444 412 0.7133 0.1379 0.7133 0.8446
No log 11.5 414 0.7283 0.1311 0.7283 0.8534
No log 11.5556 416 0.7666 0.0639 0.7666 0.8755
No log 11.6111 418 0.7649 0.1254 0.7649 0.8746
No log 11.6667 420 0.7964 0.0639 0.7964 0.8924
No log 11.7222 422 0.8082 0.0600 0.8082 0.8990
No log 11.7778 424 0.7795 -0.0113 0.7795 0.8829
No log 11.8333 426 0.8021 0.0603 0.8021 0.8956
No log 11.8889 428 0.7854 0.0444 0.7854 0.8862
No log 11.9444 430 0.7737 0.0834 0.7737 0.8796
No log 12.0 432 0.7668 0.0449 0.7668 0.8757
No log 12.0556 434 0.7672 0.0934 0.7672 0.8759
No log 12.1111 436 0.8057 0.0989 0.8057 0.8976
No log 12.1667 438 0.8034 -0.0444 0.8034 0.8963
No log 12.2222 440 0.7868 -0.0557 0.7868 0.8870
No log 12.2778 442 0.7932 0.0465 0.7932 0.8906
No log 12.3333 444 0.8094 -0.0661 0.8094 0.8997
No log 12.3889 446 0.7679 0.0432 0.7679 0.8763
No log 12.4444 448 0.7535 0.0814 0.7535 0.8680
No log 12.5 450 0.8038 -0.0355 0.8038 0.8965
No log 12.5556 452 0.7908 -0.0316 0.7908 0.8893
No log 12.6111 454 0.7418 0.0247 0.7418 0.8613
No log 12.6667 456 0.7520 0.1951 0.7520 0.8672
No log 12.7222 458 0.7502 0.0585 0.7502 0.8661
No log 12.7778 460 0.7223 0.2271 0.7223 0.8499
No log 12.8333 462 0.7750 0.1506 0.7750 0.8803
No log 12.8889 464 0.7869 0.0123 0.7869 0.8871
No log 12.9444 466 0.7708 0.1298 0.7708 0.8780
No log 13.0 468 0.7966 0.0053 0.7966 0.8925
No log 13.0556 470 0.7791 0.0412 0.7791 0.8826
No log 13.1111 472 0.7878 0.0303 0.7878 0.8876
No log 13.1667 474 0.8125 0.0146 0.8125 0.9014
No log 13.2222 476 0.8016 0.0269 0.8016 0.8953
No log 13.2778 478 0.8533 0.0799 0.8533 0.9238
No log 13.3333 480 0.9324 -0.0138 0.9324 0.9656
No log 13.3889 482 0.9452 -0.0526 0.9452 0.9722
No log 13.4444 484 0.8923 0.0277 0.8923 0.9446
No log 13.5 486 0.8014 0.0269 0.8014 0.8952
No log 13.5556 488 0.7960 0.0558 0.7960 0.8922
No log 13.6111 490 0.7829 0.0570 0.7829 0.8848
No log 13.6667 492 0.7605 0.1354 0.7605 0.8721
No log 13.7222 494 0.7756 0.0269 0.7756 0.8807
No log 13.7778 496 0.7610 0.0869 0.7610 0.8724
No log 13.8333 498 0.8179 0.0603 0.8179 0.9044
0.3298 13.8889 500 0.8521 0.1001 0.8521 0.9231
0.3298 13.9444 502 0.8966 0.0048 0.8966 0.9469
0.3298 14.0 504 0.8166 0.0999 0.8166 0.9036
0.3298 14.0556 506 0.7447 0.0869 0.7447 0.8630
0.3298 14.1111 508 0.7677 0.1506 0.7677 0.8762
0.3298 14.1667 510 0.7647 0.1047 0.7647 0.8745
0.3298 14.2222 512 0.7561 0.0791 0.7561 0.8695
0.3298 14.2778 514 0.8184 0.0996 0.8184 0.9046
0.3298 14.3333 516 0.8083 0.0993 0.8083 0.8990
0.3298 14.3889 518 0.8026 0.1393 0.8026 0.8959
0.3298 14.4444 520 0.7394 0.1778 0.7394 0.8599
0.3298 14.5 522 0.7077 0.0828 0.7077 0.8412
0.3298 14.5556 524 0.7048 0.0828 0.7048 0.8395
0.3298 14.6111 526 0.7061 0.1413 0.7061 0.8403
0.3298 14.6667 528 0.7511 0.1080 0.7511 0.8667
0.3298 14.7222 530 0.8047 0.0279 0.8047 0.8970
0.3298 14.7778 532 0.8370 0.1119 0.8370 0.9149
0.3298 14.8333 534 0.7784 0.0987 0.7784 0.8822
0.3298 14.8889 536 0.7804 0.1001 0.7804 0.8834
0.3298 14.9444 538 1.0041 0.0207 1.0041 1.0021
0.3298 15.0 540 1.1280 0.0404 1.1280 1.0621
0.3298 15.0556 542 1.0245 0.0545 1.0245 1.0122
0.3298 15.1111 544 0.8058 0.1047 0.8058 0.8976
0.3298 15.1667 546 0.7853 0.0323 0.7853 0.8862
0.3298 15.2222 548 0.7883 0.0303 0.7883 0.8878
0.3298 15.2778 550 0.7724 0.0393 0.7724 0.8789
0.3298 15.3333 552 0.7473 0.1199 0.7473 0.8645
0.3298 15.3889 554 0.8511 0.0711 0.8511 0.9226
0.3298 15.4444 556 0.9038 -0.0163 0.9038 0.9507
0.3298 15.5 558 0.8377 0.0512 0.8377 0.9152
0.3298 15.5556 560 0.7935 0.0759 0.7935 0.8908
0.3298 15.6111 562 0.7755 0.0863 0.7755 0.8806

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k13_task3_organization

Finetuned
(4023)
this model