ArabicNewSplits6_FineTuningAraBERT_run1_AugV5_k10_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0186
  • Qwk: 0.1367
  • Mse: 1.0186
  • Rmse: 1.0093

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0417 2 3.3525 -0.0227 3.3525 1.8310
No log 0.0833 4 1.6720 -0.0070 1.6720 1.2931
No log 0.125 6 1.3667 -0.0035 1.3667 1.1690
No log 0.1667 8 0.8938 0.0617 0.8938 0.9454
No log 0.2083 10 0.5799 0.0303 0.5799 0.7615
No log 0.25 12 0.5602 0.0388 0.5602 0.7485
No log 0.2917 14 0.8121 0.2000 0.8121 0.9012
No log 0.3333 16 0.9018 0.1486 0.9018 0.9496
No log 0.375 18 1.1279 0.0 1.1279 1.0620
No log 0.4167 20 1.2348 0.0 1.2348 1.1112
No log 0.4583 22 1.0177 0.0 1.0177 1.0088
No log 0.5 24 0.7029 0.0476 0.7029 0.8384
No log 0.5417 26 0.7413 0.0 0.7413 0.8610
No log 0.5833 28 0.7339 0.1186 0.7339 0.8567
No log 0.625 30 0.6943 0.1724 0.6943 0.8332
No log 0.6667 32 0.7218 0.1398 0.7218 0.8496
No log 0.7083 34 0.6476 0.0556 0.6476 0.8047
No log 0.75 36 0.5980 0.1111 0.5980 0.7733
No log 0.7917 38 0.6386 0.2308 0.6386 0.7991
No log 0.8333 40 0.7893 0.1111 0.7893 0.8884
No log 0.875 42 0.8834 0.1111 0.8834 0.9399
No log 0.9167 44 1.0416 0.0388 1.0416 1.0206
No log 0.9583 46 0.8346 0.1169 0.8346 0.9136
No log 1.0 48 0.5776 0.2533 0.5776 0.7600
No log 1.0417 50 0.5750 0.0303 0.5750 0.7583
No log 1.0833 52 0.6315 0.0 0.6315 0.7947
No log 1.125 54 0.6494 -0.0068 0.6494 0.8058
No log 1.1667 56 0.7017 0.0877 0.7017 0.8377
No log 1.2083 58 0.7271 0.0734 0.7271 0.8527
No log 1.25 60 0.8363 0.1590 0.8363 0.9145
No log 1.2917 62 0.6130 -0.0370 0.6130 0.7830
No log 1.3333 64 0.6941 0.1895 0.6941 0.8331
No log 1.375 66 0.6018 0.1176 0.6018 0.7757
No log 1.4167 68 0.6484 0.0692 0.6484 0.8053
No log 1.4583 70 1.2966 0.1081 1.2966 1.1387
No log 1.5 72 1.7201 -0.0157 1.7201 1.3115
No log 1.5417 74 1.3758 0.0426 1.3758 1.1730
No log 1.5833 76 0.9362 0.0745 0.9362 0.9676
No log 1.625 78 0.6452 0.2485 0.6452 0.8032
No log 1.6667 80 0.5814 0.1467 0.5814 0.7625
No log 1.7083 82 0.6988 0.1549 0.6988 0.8359
No log 1.75 84 0.6914 0.2077 0.6914 0.8315
No log 1.7917 86 0.6066 0.1724 0.6066 0.7788
No log 1.8333 88 0.5965 0.1111 0.5965 0.7723
No log 1.875 90 0.6536 0.0123 0.6536 0.8084
No log 1.9167 92 0.6737 0.0545 0.6737 0.8208
No log 1.9583 94 0.7415 0.1913 0.7415 0.8611
No log 2.0 96 0.6461 0.1429 0.6461 0.8038
No log 2.0417 98 0.5952 0.1533 0.5952 0.7715
No log 2.0833 100 0.6121 0.0448 0.6121 0.7823
No log 2.125 102 0.6257 0.0448 0.6257 0.7910
No log 2.1667 104 0.6399 -0.0963 0.6399 0.7999
No log 2.2083 106 0.7075 0.1373 0.7075 0.8411
No log 2.25 108 0.6858 0.2000 0.6858 0.8282
No log 2.2917 110 0.6567 0.0303 0.6567 0.8104
No log 2.3333 112 0.6665 -0.0076 0.6665 0.8164
No log 2.375 114 0.6590 0.0067 0.6590 0.8118
No log 2.4167 116 0.7817 0.1837 0.7817 0.8842
No log 2.4583 118 0.7542 0.1648 0.7542 0.8684
No log 2.5 120 0.6802 0.1732 0.6802 0.8248
No log 2.5417 122 0.6168 0.1373 0.6168 0.7853
No log 2.5833 124 0.6486 0.2970 0.6486 0.8053
No log 2.625 126 0.6079 0.3023 0.6079 0.7797
No log 2.6667 128 0.7618 0.0800 0.7618 0.8728
No log 2.7083 130 0.8124 0.0707 0.8124 0.9013
No log 2.75 132 0.6538 0.2832 0.6538 0.8086
No log 2.7917 134 0.6252 0.2771 0.6252 0.7907
No log 2.8333 136 0.6323 0.2771 0.6323 0.7952
No log 2.875 138 0.6483 0.2381 0.6483 0.8052
No log 2.9167 140 0.6674 0.2727 0.6674 0.8170
No log 2.9583 142 0.7114 0.1841 0.7114 0.8434
No log 3.0 144 0.7347 0.2083 0.7347 0.8572
No log 3.0417 146 0.7632 0.2381 0.7632 0.8736
No log 3.0833 148 0.8188 0.1289 0.8188 0.9049
No log 3.125 150 0.8089 0.2607 0.8089 0.8994
No log 3.1667 152 1.0577 0.1471 1.0577 1.0284
No log 3.2083 154 0.9718 0.1938 0.9718 0.9858
No log 3.25 156 0.9669 0.1319 0.9669 0.9833
No log 3.2917 158 1.3813 0.1419 1.3813 1.1753
No log 3.3333 160 1.3367 0.1126 1.3367 1.1561
No log 3.375 162 0.9392 0.2000 0.9392 0.9691
No log 3.4167 164 0.7937 0.2000 0.7937 0.8909
No log 3.4583 166 0.7673 0.2410 0.7673 0.8760
No log 3.5 168 0.7323 0.1456 0.7323 0.8557
No log 3.5417 170 0.8292 0.1786 0.8292 0.9106
No log 3.5833 172 0.9134 0.1605 0.9134 0.9557
No log 3.625 174 0.8526 0.1803 0.8526 0.9233
No log 3.6667 176 0.7346 0.2227 0.7346 0.8571
No log 3.7083 178 0.6918 0.3231 0.6918 0.8317
No log 3.75 180 0.7031 0.2990 0.7031 0.8385
No log 3.7917 182 0.7158 0.2990 0.7158 0.8461
No log 3.8333 184 0.7190 0.2077 0.7190 0.8479
No log 3.875 186 0.7244 0.2637 0.7244 0.8511
No log 3.9167 188 0.8337 0.2000 0.8337 0.9131
No log 3.9583 190 0.9525 0.1220 0.9525 0.9760
No log 4.0 192 0.8699 0.2070 0.8699 0.9327
No log 4.0417 194 0.7639 0.1852 0.7639 0.8740
No log 4.0833 196 0.7886 0.1781 0.7886 0.8881
No log 4.125 198 0.9036 0.2000 0.9036 0.9506
No log 4.1667 200 0.9947 0.0977 0.9947 0.9973
No log 4.2083 202 0.9305 0.0769 0.9305 0.9646
No log 4.25 204 0.9800 0.1545 0.9800 0.9900
No log 4.2917 206 0.7859 0.1193 0.7859 0.8865
No log 4.3333 208 0.7344 0.2077 0.7344 0.8570
No log 4.375 210 0.6996 0.2245 0.6996 0.8364
No log 4.4167 212 0.7432 0.1429 0.7432 0.8621
No log 4.4583 214 0.8611 0.1549 0.8611 0.9280
No log 4.5 216 0.7811 0.0680 0.7811 0.8838
No log 4.5417 218 0.6790 0.2787 0.6790 0.8240
No log 4.5833 220 0.7227 0.2239 0.7227 0.8501
No log 4.625 222 0.7409 0.2381 0.7409 0.8607
No log 4.6667 224 0.8168 0.2000 0.8168 0.9037
No log 4.7083 226 0.9878 0.0958 0.9878 0.9939
No log 4.75 228 0.9298 0.2258 0.9298 0.9642
No log 4.7917 230 0.8232 0.2711 0.8232 0.9073
No log 4.8333 232 0.8593 0.1504 0.8593 0.9270
No log 4.875 234 0.9944 0.0938 0.9944 0.9972
No log 4.9167 236 1.1595 0.1111 1.1595 1.0768
No log 4.9583 238 1.2365 0.1399 1.2365 1.1120
No log 5.0 240 0.9882 0.1496 0.9882 0.9941
No log 5.0417 242 0.7857 0.3153 0.7857 0.8864
No log 5.0833 244 0.9106 0.2000 0.9106 0.9543
No log 5.125 246 0.8145 0.3363 0.8145 0.9025
No log 5.1667 248 0.8184 0.2217 0.8184 0.9046
No log 5.2083 250 1.1260 0.1079 1.1260 1.0611
No log 5.25 252 1.1200 0.1143 1.1200 1.0583
No log 5.2917 254 0.9843 0.1027 0.9843 0.9921
No log 5.3333 256 0.8231 0.1786 0.8231 0.9072
No log 5.375 258 0.7433 0.2000 0.7433 0.8622
No log 5.4167 260 0.7335 0.2000 0.7335 0.8565
No log 5.4583 262 0.8506 0.1515 0.8506 0.9223
No log 5.5 264 0.9199 0.1597 0.9199 0.9591
No log 5.5417 266 0.8547 0.1535 0.8547 0.9245
No log 5.5833 268 0.7045 0.3171 0.7045 0.8393
No log 5.625 270 0.6800 0.3077 0.6800 0.8246
No log 5.6667 272 0.7282 0.2646 0.7282 0.8534
No log 5.7083 274 0.7413 0.2579 0.7413 0.8610
No log 5.75 276 0.8379 0.1790 0.8379 0.9154
No log 5.7917 278 1.0095 0.0988 1.0095 1.0048
No log 5.8333 280 1.0492 0.1045 1.0492 1.0243
No log 5.875 282 1.0411 0.0746 1.0411 1.0204
No log 5.9167 284 0.9603 0.1235 0.9603 0.9799
No log 5.9583 286 0.9764 0.0938 0.9764 0.9881
No log 6.0 288 0.9063 0.0492 0.9063 0.9520
No log 6.0417 290 0.7905 0.2432 0.7905 0.8891
No log 6.0833 292 0.7658 0.2000 0.7658 0.8751
No log 6.125 294 0.8355 0.2520 0.8355 0.9141
No log 6.1667 296 0.9638 0.1385 0.9638 0.9817
No log 6.2083 298 1.0381 0.0836 1.0381 1.0189
No log 6.25 300 0.9247 0.1937 0.9247 0.9616
No log 6.2917 302 0.8709 0.2253 0.8709 0.9332
No log 6.3333 304 0.7465 0.1691 0.7465 0.8640
No log 6.375 306 0.7274 0.1691 0.7274 0.8529
No log 6.4167 308 0.8109 0.1660 0.8109 0.9005
No log 6.4583 310 0.9392 0.1165 0.9392 0.9691
No log 6.5 312 0.9562 0.1181 0.9562 0.9778
No log 6.5417 314 0.8330 0.2000 0.8330 0.9127
No log 6.5833 316 0.7881 0.1855 0.7881 0.8877
No log 6.625 318 0.7502 0.1698 0.7502 0.8662
No log 6.6667 320 0.7481 0.1636 0.7481 0.8649
No log 6.7083 322 0.7856 0.2000 0.7856 0.8863
No log 6.75 324 0.8206 0.1667 0.8206 0.9058
No log 6.7917 326 0.8573 0.1417 0.8573 0.9259
No log 6.8333 328 0.9349 0.2868 0.9349 0.9669
No log 6.875 330 1.0059 0.1815 1.0059 1.0029
No log 6.9167 332 0.9900 0.2302 0.9900 0.9950
No log 6.9583 334 0.8834 0.1535 0.8834 0.9399
No log 7.0 336 0.8740 0.1535 0.8740 0.9349
No log 7.0417 338 0.9116 0.1937 0.9116 0.9548
No log 7.0833 340 0.9577 0.1940 0.9577 0.9786
No log 7.125 342 0.8830 0.1613 0.8830 0.9397
No log 7.1667 344 0.7854 0.2423 0.7854 0.8863
No log 7.2083 346 0.7681 0.2423 0.7681 0.8764
No log 7.25 348 0.8083 0.1515 0.8083 0.8990
No log 7.2917 350 0.8550 0.1276 0.8550 0.9247
No log 7.3333 352 0.9295 0.2062 0.9295 0.9641
No log 7.375 354 0.9111 0.2062 0.9111 0.9545
No log 7.4167 356 0.8211 0.1211 0.8211 0.9062
No log 7.4583 358 0.7755 0.2143 0.7755 0.8806
No log 7.5 360 0.8027 0.1790 0.8027 0.8960
No log 7.5417 362 0.8535 0.0588 0.8535 0.9238
No log 7.5833 364 0.9352 0.0745 0.9352 0.9670
No log 7.625 366 0.9371 0.0745 0.9371 0.9680
No log 7.6667 368 0.9608 0.0769 0.9608 0.9802
No log 7.7083 370 0.9786 0.1062 0.9786 0.9892
No log 7.75 372 0.9630 0.1027 0.9630 0.9813
No log 7.7917 374 0.9847 0.1343 0.9847 0.9923
No log 7.8333 376 0.9957 0.1062 0.9957 0.9979
No log 7.875 378 1.0001 0.1062 1.0001 1.0000
No log 7.9167 380 0.9485 0.1008 0.9485 0.9739
No log 7.9583 382 0.9331 0.1235 0.9331 0.9660
No log 8.0 384 0.9503 0.0938 0.9503 0.9748
No log 8.0417 386 0.9881 0.1587 0.9881 0.9941
No log 8.0833 388 1.0568 0.1667 1.0568 1.0280
No log 8.125 390 1.0872 0.1667 1.0872 1.0427
No log 8.1667 392 1.0273 0.1667 1.0273 1.0136
No log 8.2083 394 0.9421 0.1197 0.9421 0.9706
No log 8.25 396 0.9144 0.1429 0.9144 0.9562
No log 8.2917 398 0.9159 0.1429 0.9159 0.9570
No log 8.3333 400 0.9227 0.1181 0.9227 0.9606
No log 8.375 402 0.9233 0.0588 0.9233 0.9609
No log 8.4167 404 0.9025 0.0558 0.9025 0.9500
No log 8.4583 406 0.9100 0.0609 0.9100 0.9540
No log 8.5 408 0.9083 0.0638 0.9083 0.9530
No log 8.5417 410 0.8919 0.0609 0.8919 0.9444
No log 8.5833 412 0.9142 0.0367 0.9142 0.9561
No log 8.625 414 0.9771 0.0720 0.9771 0.9885
No log 8.6667 416 1.0142 0.1059 1.0142 1.0071
No log 8.7083 418 1.0341 0.1450 1.0341 1.0169
No log 8.75 420 1.0493 0.1450 1.0493 1.0244
No log 8.7917 422 1.0232 0.1094 1.0232 1.0116
No log 8.8333 424 0.9973 0.1059 0.9973 0.9986
No log 8.875 426 0.9953 0.1059 0.9953 0.9977
No log 8.9167 428 0.9531 0.0720 0.9531 0.9762
No log 8.9583 430 0.9367 0.0367 0.9367 0.9678
No log 9.0 432 0.9228 0.0323 0.9228 0.9606
No log 9.0417 434 0.9285 0.0323 0.9285 0.9636
No log 9.0833 436 0.9383 0.0672 0.9383 0.9686
No log 9.125 438 0.9271 0.0323 0.9271 0.9629
No log 9.1667 440 0.9277 0.0323 0.9277 0.9632
No log 9.2083 442 0.9476 0.0672 0.9476 0.9735
No log 9.25 444 0.9720 0.0720 0.9720 0.9859
No log 9.2917 446 0.9784 0.1027 0.9784 0.9892
No log 9.3333 448 0.9945 0.1027 0.9945 0.9973
No log 9.375 450 1.0053 0.1027 1.0053 1.0027
No log 9.4167 452 1.0092 0.1343 1.0092 1.0046
No log 9.4583 454 1.0219 0.1367 1.0219 1.0109
No log 9.5 456 1.0250 0.1367 1.0250 1.0124
No log 9.5417 458 1.0175 0.1062 1.0175 1.0087
No log 9.5833 460 1.0131 0.1062 1.0131 1.0065
No log 9.625 462 1.0186 0.1367 1.0186 1.0093
No log 9.6667 464 1.0262 0.1367 1.0262 1.0130
No log 9.7083 466 1.0257 0.1367 1.0257 1.0128
No log 9.75 468 1.0284 0.1367 1.0284 1.0141
No log 9.7917 470 1.0255 0.1367 1.0255 1.0127
No log 9.8333 472 1.0199 0.1367 1.0199 1.0099
No log 9.875 474 1.0181 0.1367 1.0181 1.0090
No log 9.9167 476 1.0193 0.1367 1.0193 1.0096
No log 9.9583 478 1.0193 0.1367 1.0193 1.0096
No log 10.0 480 1.0186 0.1367 1.0186 1.0093

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_FineTuningAraBERT_run1_AugV5_k10_task3_organization

Finetuned
(4023)
this model