ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k14_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9746
  • Qwk: 0.6467
  • Mse: 0.9746
  • Rmse: 0.9872

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0303 2 6.8908 0.0242 6.8908 2.6250
No log 0.0606 4 4.5354 0.0706 4.5354 2.1296
No log 0.0909 6 3.3376 0.0222 3.3376 1.8269
No log 0.1212 8 2.4049 0.1159 2.4049 1.5508
No log 0.1515 10 1.8071 0.1897 1.8071 1.3443
No log 0.1818 12 1.7995 0.2124 1.7995 1.3415
No log 0.2121 14 2.0644 0.1653 2.0644 1.4368
No log 0.2424 16 1.9296 0.1653 1.9296 1.3891
No log 0.2727 18 1.7741 0.2222 1.7741 1.3320
No log 0.3030 20 1.7429 0.2281 1.7429 1.3202
No log 0.3333 22 1.5268 0.2321 1.5268 1.2356
No log 0.3636 24 1.4701 0.2679 1.4701 1.2125
No log 0.3939 26 1.5620 0.2951 1.5620 1.2498
No log 0.4242 28 1.7486 0.3053 1.7486 1.3223
No log 0.4545 30 1.6992 0.3053 1.6992 1.3035
No log 0.4848 32 2.1199 0.2313 2.1199 1.4560
No log 0.5152 34 2.5852 0.2038 2.5852 1.6079
No log 0.5455 36 3.2983 0.1224 3.2983 1.8161
No log 0.5758 38 3.7765 0.1233 3.7765 1.9433
No log 0.6061 40 3.3168 0.2056 3.3168 1.8212
No log 0.6364 42 1.9289 0.3905 1.9289 1.3889
No log 0.6667 44 1.1937 0.5547 1.1937 1.0926
No log 0.6970 46 1.0719 0.6853 1.0719 1.0353
No log 0.7273 48 1.7753 0.3841 1.7753 1.3324
No log 0.7576 50 2.9654 0.2151 2.9654 1.7220
No log 0.7879 52 4.0325 0.1399 4.0325 2.0081
No log 0.8182 54 3.8022 0.1502 3.8022 1.9499
No log 0.8485 56 3.0613 0.1618 3.0613 1.7497
No log 0.8788 58 2.3358 0.0667 2.3358 1.5283
No log 0.9091 60 1.8832 0.2794 1.8832 1.3723
No log 0.9394 62 1.5164 0.4603 1.5164 1.2314
No log 0.9697 64 1.1223 0.4576 1.1223 1.0594
No log 1.0 66 1.1836 0.5691 1.1836 1.0879
No log 1.0303 68 1.5672 0.3009 1.5672 1.2519
No log 1.0606 70 1.6143 0.2478 1.6143 1.2706
No log 1.0909 72 1.3973 0.2807 1.3973 1.1821
No log 1.1212 74 1.2330 0.4034 1.2330 1.1104
No log 1.1515 76 1.4574 0.4328 1.4574 1.2072
No log 1.1818 78 1.9528 0.3472 1.9528 1.3974
No log 1.2121 80 2.0405 0.25 2.0405 1.4285
No log 1.2424 82 1.8824 0.3194 1.8824 1.3720
No log 1.2727 84 1.8228 0.3217 1.8228 1.3501
No log 1.3030 86 1.8121 0.4028 1.8121 1.3461
No log 1.3333 88 1.6244 0.3946 1.6244 1.2745
No log 1.3636 90 1.6159 0.4416 1.6159 1.2712
No log 1.3939 92 1.6071 0.4941 1.6071 1.2677
No log 1.4242 94 1.6344 0.4941 1.6344 1.2785
No log 1.4545 96 1.3679 0.5952 1.3679 1.1696
No log 1.4848 98 1.1322 0.6203 1.1322 1.0640
No log 1.5152 100 0.9217 0.6957 0.9217 0.9600
No log 1.5455 102 0.8273 0.7483 0.8273 0.9095
No log 1.5758 104 0.8352 0.7432 0.8352 0.9139
No log 1.6061 106 0.8615 0.6897 0.8615 0.9282
No log 1.6364 108 0.8356 0.6897 0.8356 0.9141
No log 1.6667 110 0.8411 0.6713 0.8411 0.9171
No log 1.6970 112 1.0312 0.6027 1.0312 1.0155
No log 1.7273 114 1.1155 0.5974 1.1155 1.0562
No log 1.7576 116 1.0813 0.6220 1.0813 1.0399
No log 1.7879 118 0.9012 0.6667 0.9012 0.9493
No log 1.8182 120 0.8765 0.7182 0.8765 0.9362
No log 1.8485 122 0.9341 0.7053 0.9341 0.9665
No log 1.8788 124 0.8820 0.7320 0.8820 0.9391
No log 1.9091 126 0.9001 0.7539 0.9001 0.9487
No log 1.9394 128 0.9241 0.7166 0.9241 0.9613
No log 1.9697 130 0.9034 0.7174 0.9034 0.9505
No log 2.0 132 0.8439 0.7419 0.8439 0.9187
No log 2.0303 134 0.7973 0.7630 0.7973 0.8929
No log 2.0606 136 0.7407 0.7907 0.7407 0.8606
No log 2.0909 138 0.7237 0.7778 0.7237 0.8507
No log 2.1212 140 0.6594 0.7632 0.6594 0.8120
No log 2.1515 142 0.6601 0.7451 0.6601 0.8124
No log 2.1818 144 0.6332 0.7712 0.6332 0.7957
No log 2.2121 146 0.6152 0.7712 0.6152 0.7843
No log 2.2424 148 0.5994 0.7895 0.5994 0.7742
No log 2.2727 150 0.6194 0.7922 0.6194 0.7870
No log 2.3030 152 0.6267 0.8101 0.6267 0.7916
No log 2.3333 154 0.6270 0.7925 0.6270 0.7918
No log 2.3636 156 0.6652 0.7821 0.6652 0.8156
No log 2.3939 158 0.7158 0.7564 0.7158 0.8461
No log 2.4242 160 0.8017 0.75 0.8017 0.8954
No log 2.4545 162 1.1812 0.6429 1.1812 1.0868
No log 2.4848 164 1.2840 0.6257 1.2840 1.1331
No log 2.5152 166 1.1372 0.6265 1.1372 1.0664
No log 2.5455 168 0.8742 0.7215 0.8742 0.9350
No log 2.5758 170 0.7399 0.7799 0.7399 0.8602
No log 2.6061 172 0.7141 0.7922 0.7141 0.8450
No log 2.6364 174 0.6790 0.8199 0.6790 0.8240
No log 2.6667 176 0.6836 0.8272 0.6836 0.8268
No log 2.6970 178 0.6836 0.8272 0.6836 0.8268
No log 2.7273 180 0.7398 0.7607 0.7398 0.8601
No log 2.7576 182 0.9107 0.6832 0.9107 0.9543
No log 2.7879 184 1.0208 0.6667 1.0208 1.0103
No log 2.8182 186 0.9050 0.6752 0.9050 0.9513
No log 2.8485 188 0.7362 0.7273 0.7362 0.8580
No log 2.8788 190 0.7379 0.7413 0.7379 0.8590
No log 2.9091 192 0.7385 0.7619 0.7385 0.8594
No log 2.9394 194 0.7008 0.7763 0.7008 0.8371
No log 2.9697 196 0.7161 0.7517 0.7161 0.8463
No log 3.0 198 0.6772 0.7651 0.6772 0.8229
No log 3.0303 200 0.6655 0.7792 0.6655 0.8158
No log 3.0606 202 0.8075 0.7051 0.8075 0.8986
No log 3.0909 204 0.8949 0.6575 0.8949 0.9460
No log 3.1212 206 0.8852 0.5839 0.8852 0.9409
No log 3.1515 208 0.8608 0.5899 0.8608 0.9278
No log 3.1818 210 0.7623 0.6528 0.7623 0.8731
No log 3.2121 212 0.6641 0.7333 0.6641 0.8149
No log 3.2424 214 0.6531 0.7879 0.6531 0.8081
No log 3.2727 216 0.7096 0.7574 0.7096 0.8424
No log 3.3030 218 0.6386 0.8214 0.6386 0.7991
No log 3.3333 220 0.5440 0.8221 0.5440 0.7375
No log 3.3636 222 0.5714 0.7974 0.5714 0.7559
No log 3.3939 224 0.5821 0.7867 0.5821 0.7629
No log 3.4242 226 0.6882 0.7333 0.6882 0.8296
No log 3.4545 228 0.8042 0.7308 0.8042 0.8968
No log 3.4848 230 0.7106 0.7333 0.7106 0.8430
No log 3.5152 232 0.6267 0.7682 0.6267 0.7917
No log 3.5455 234 0.6503 0.7517 0.6503 0.8064
No log 3.5758 236 0.6472 0.7815 0.6472 0.8045
No log 3.6061 238 0.6565 0.7799 0.6565 0.8102
No log 3.6364 240 0.7152 0.7722 0.7152 0.8457
No log 3.6667 242 0.7528 0.7211 0.7528 0.8677
No log 3.6970 244 0.7760 0.7361 0.7760 0.8809
No log 3.7273 246 0.7794 0.7429 0.7794 0.8828
No log 3.7576 248 0.7264 0.7448 0.7264 0.8523
No log 3.7879 250 0.6345 0.7632 0.6345 0.7966
No log 3.8182 252 0.6746 0.7516 0.6746 0.8213
No log 3.8485 254 0.7345 0.7329 0.7345 0.8570
No log 3.8788 256 0.6425 0.7826 0.6425 0.8015
No log 3.9091 258 0.5754 0.7975 0.5754 0.7586
No log 3.9394 260 0.6197 0.7925 0.6197 0.7872
No log 3.9697 262 0.6225 0.8 0.6225 0.7890
No log 4.0 264 0.6157 0.7975 0.6157 0.7846
No log 4.0303 266 0.6766 0.8114 0.6766 0.8225
No log 4.0606 268 0.7748 0.7771 0.7748 0.8802
No log 4.0909 270 0.7664 0.7886 0.7664 0.8754
No log 4.1212 272 0.7081 0.8072 0.7081 0.8415
No log 4.1515 274 0.6922 0.8049 0.6922 0.8320
No log 4.1818 276 0.6724 0.8075 0.6724 0.8200
No log 4.2121 278 0.6589 0.7843 0.6589 0.8117
No log 4.2424 280 0.6703 0.7843 0.6703 0.8187
No log 4.2727 282 0.7219 0.7517 0.7219 0.8497
No log 4.3030 284 0.7709 0.6667 0.7709 0.8780
No log 4.3333 286 0.7317 0.6759 0.7317 0.8554
No log 4.3636 288 0.6386 0.7755 0.6386 0.7991
No log 4.3939 290 0.6252 0.7919 0.6252 0.7907
No log 4.4242 292 0.6491 0.7534 0.6491 0.8057
No log 4.4545 294 0.6249 0.7867 0.6249 0.7905
No log 4.4848 296 0.5833 0.7919 0.5833 0.7637
No log 4.5152 298 0.5935 0.7838 0.5935 0.7704
No log 4.5455 300 0.6071 0.7838 0.6071 0.7792
No log 4.5758 302 0.5962 0.7919 0.5962 0.7721
No log 4.6061 304 0.6010 0.7712 0.6010 0.7753
No log 4.6364 306 0.6688 0.7904 0.6688 0.8178
No log 4.6667 308 0.7083 0.7657 0.7083 0.8416
No log 4.6970 310 0.7001 0.7816 0.7001 0.8367
No log 4.7273 312 0.6283 0.7882 0.6283 0.7927
No log 4.7576 314 0.7008 0.7771 0.7008 0.8371
No log 4.7879 316 0.7826 0.7582 0.7826 0.8846
No log 4.8182 318 0.6949 0.7771 0.6949 0.8336
No log 4.8485 320 0.7039 0.7654 0.7039 0.8390
No log 4.8788 322 0.9070 0.7024 0.9070 0.9524
No log 4.9091 324 1.0419 0.6424 1.0419 1.0208
No log 4.9394 326 0.8772 0.7 0.8772 0.9366
No log 4.9697 328 0.6844 0.7692 0.6844 0.8273
No log 5.0 330 0.7045 0.7324 0.7045 0.8393
No log 5.0303 332 0.7662 0.7143 0.7662 0.8753
No log 5.0606 334 0.7903 0.7273 0.7903 0.8890
No log 5.0909 336 0.8127 0.6944 0.8127 0.9015
No log 5.1212 338 0.7839 0.72 0.7839 0.8854
No log 5.1515 340 0.7176 0.7248 0.7176 0.8471
No log 5.1818 342 0.6213 0.7397 0.6213 0.7882
No log 5.2121 344 0.6161 0.7586 0.6161 0.7850
No log 5.2424 346 0.6718 0.7285 0.6718 0.8197
No log 5.2727 348 0.6827 0.7285 0.6827 0.8262
No log 5.3030 350 0.6115 0.7552 0.6115 0.7820
No log 5.3333 352 0.6293 0.7310 0.6293 0.7933
No log 5.3636 354 0.7614 0.6849 0.7614 0.8726
No log 5.3939 356 0.7840 0.7114 0.7840 0.8854
No log 5.4242 358 0.6637 0.7211 0.6637 0.8147
No log 5.4545 360 0.5998 0.7947 0.5998 0.7745
No log 5.4848 362 0.6003 0.7947 0.6003 0.7748
No log 5.5152 364 0.5935 0.7895 0.5935 0.7704
No log 5.5455 366 0.6085 0.7792 0.6085 0.7801
No log 5.5758 368 0.5997 0.8025 0.5997 0.7744
No log 5.6061 370 0.6079 0.8095 0.6079 0.7797
No log 5.6364 372 0.5999 0.8 0.5999 0.7745
No log 5.6667 374 0.6024 0.8166 0.6024 0.7762
No log 5.6970 376 0.5852 0.8025 0.5852 0.7650
No log 5.7273 378 0.5814 0.8 0.5814 0.7625
No log 5.7576 380 0.5944 0.7895 0.5944 0.7710
No log 5.7879 382 0.6028 0.7895 0.6028 0.7764
No log 5.8182 384 0.6062 0.7895 0.6062 0.7786
No log 5.8485 386 0.6445 0.8098 0.6445 0.8028
No log 5.8788 388 0.7475 0.7826 0.7475 0.8646
No log 5.9091 390 0.8040 0.7673 0.8040 0.8967
No log 5.9394 392 0.8417 0.7468 0.8417 0.9174
No log 5.9697 394 0.7654 0.7143 0.7654 0.8748
No log 6.0 396 0.7230 0.7152 0.7230 0.8503
No log 6.0303 398 0.6627 0.7651 0.6627 0.8140
No log 6.0606 400 0.6498 0.7947 0.6498 0.8061
No log 6.0909 402 0.6376 0.8 0.6376 0.7985
No log 6.1212 404 0.6594 0.7564 0.6594 0.8120
No log 6.1515 406 0.7072 0.7389 0.7072 0.8410
No log 6.1818 408 0.7174 0.7297 0.7174 0.8470
No log 6.2121 410 0.7305 0.7172 0.7305 0.8547
No log 6.2424 412 0.7142 0.7172 0.7142 0.8451
No log 6.2727 414 0.7245 0.7397 0.7245 0.8512
No log 6.3030 416 0.7851 0.7114 0.7851 0.8860
No log 6.3333 418 0.7980 0.7468 0.7980 0.8933
No log 6.3636 420 0.7849 0.7375 0.7849 0.8859
No log 6.3939 422 0.6814 0.7595 0.6814 0.8255
No log 6.4242 424 0.6099 0.7815 0.6099 0.7810
No log 6.4545 426 0.6239 0.7703 0.6239 0.7898
No log 6.4848 428 0.6561 0.7639 0.6561 0.8100
No log 6.5152 430 0.6888 0.7606 0.6888 0.8300
No log 6.5455 432 0.7172 0.7445 0.7172 0.8469
No log 6.5758 434 0.7310 0.7338 0.7310 0.8550
No log 6.6061 436 0.6990 0.7429 0.6990 0.8361
No log 6.6364 438 0.6487 0.7448 0.6487 0.8054
No log 6.6667 440 0.6412 0.7383 0.6412 0.8008
No log 6.6970 442 0.6271 0.76 0.6271 0.7919
No log 6.7273 444 0.6301 0.7682 0.6301 0.7938
No log 6.7576 446 0.6255 0.7763 0.6255 0.7909
No log 6.7879 448 0.6316 0.7712 0.6316 0.7948
No log 6.8182 450 0.6723 0.7871 0.6723 0.8199
No log 6.8485 452 0.7781 0.7205 0.7781 0.8821
No log 6.8788 454 0.9241 0.6792 0.9241 0.9613
No log 6.9091 456 0.9128 0.6792 0.9128 0.9554
No log 6.9394 458 0.7553 0.7421 0.7553 0.8691
No log 6.9697 460 0.6098 0.7662 0.6098 0.7809
No log 7.0 462 0.5906 0.7632 0.5906 0.7685
No log 7.0303 464 0.5914 0.7867 0.5914 0.7691
No log 7.0606 466 0.6091 0.7755 0.6091 0.7804
No log 7.0909 468 0.6306 0.7619 0.6306 0.7941
No log 7.1212 470 0.6895 0.6901 0.6895 0.8304
No log 7.1515 472 0.7214 0.6806 0.7214 0.8494
No log 7.1818 474 0.6877 0.6901 0.6877 0.8293
No log 7.2121 476 0.6325 0.7692 0.6325 0.7953
No log 7.2424 478 0.6210 0.7518 0.6210 0.7880
No log 7.2727 480 0.6017 0.7660 0.6017 0.7757
No log 7.3030 482 0.6055 0.7703 0.6055 0.7782
No log 7.3333 484 0.6670 0.7260 0.6670 0.8167
No log 7.3636 486 0.6426 0.7619 0.6426 0.8016
No log 7.3939 488 0.6388 0.7571 0.6388 0.7992
No log 7.4242 490 0.6856 0.7246 0.6856 0.8280
No log 7.4545 492 0.7344 0.6866 0.7344 0.8570
No log 7.4848 494 0.7106 0.7111 0.7106 0.8430
No log 7.5152 496 0.6676 0.7445 0.6676 0.8171
No log 7.5455 498 0.7176 0.7183 0.7176 0.8471
0.4448 7.5758 500 0.7622 0.6761 0.7622 0.8730
0.4448 7.6061 502 0.7463 0.6957 0.7463 0.8639
0.4448 7.6364 504 0.7213 0.7153 0.7213 0.8493
0.4448 7.6667 506 0.7015 0.7482 0.7015 0.8376
0.4448 7.6970 508 0.6707 0.7482 0.6707 0.8190
0.4448 7.7273 510 0.6276 0.7660 0.6276 0.7922
0.4448 7.7576 512 0.5958 0.7724 0.5958 0.7719
0.4448 7.7879 514 0.5869 0.7815 0.5869 0.7661
0.4448 7.8182 516 0.5974 0.7815 0.5974 0.7729
0.4448 7.8485 518 0.6245 0.7733 0.6245 0.7903
0.4448 7.8788 520 0.6797 0.7383 0.6797 0.8245
0.4448 7.9091 522 0.7002 0.7297 0.7002 0.8368
0.4448 7.9394 524 0.7036 0.7383 0.7036 0.8388
0.4448 7.9697 526 0.6924 0.7310 0.6924 0.8321
0.4448 8.0 528 0.6886 0.7376 0.6886 0.8298
0.4448 8.0303 530 0.7176 0.6950 0.7176 0.8471
0.4448 8.0606 532 0.7949 0.6525 0.7949 0.8916
0.4448 8.0909 534 0.8728 0.6536 0.8728 0.9342
0.4448 8.1212 536 0.8689 0.6790 0.8689 0.9321
0.4448 8.1515 538 0.7407 0.6875 0.7407 0.8606
0.4448 8.1818 540 0.6646 0.7949 0.6646 0.8152
0.4448 8.2121 542 0.6911 0.7831 0.6911 0.8313
0.4448 8.2424 544 0.8001 0.6867 0.8001 0.8945
0.4448 8.2727 546 0.8172 0.6788 0.8172 0.9040
0.4448 8.3030 548 0.8225 0.6584 0.8225 0.9069
0.4448 8.3333 550 0.7585 0.6797 0.7585 0.8709
0.4448 8.3636 552 0.7644 0.6846 0.7644 0.8743
0.4448 8.3939 554 0.8631 0.6503 0.8631 0.9290
0.4448 8.4242 556 1.0257 0.6587 1.0257 1.0128
0.4448 8.4545 558 1.2956 0.6118 1.2956 1.1382
0.4448 8.4848 560 1.2643 0.6118 1.2643 1.1244
0.4448 8.5152 562 0.9746 0.6467 0.9746 0.9872

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run1_AugV5_k14_task1_organization

Finetuned
(4019)
this model