ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k12_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7846
  • Qwk: 0.6924
  • Mse: 0.7846
  • Rmse: 0.8858

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0408 2 2.1887 0.0210 2.1887 1.4794
No log 0.0816 4 1.6490 0.0595 1.6490 1.2841
No log 0.1224 6 1.7067 -0.0323 1.7067 1.3064
No log 0.1633 8 1.7278 0.1587 1.7278 1.3145
No log 0.2041 10 1.8025 0.2819 1.8025 1.3426
No log 0.2449 12 1.7793 0.3237 1.7793 1.3339
No log 0.2857 14 1.5213 0.3105 1.5213 1.2334
No log 0.3265 16 1.3226 0.2116 1.3226 1.1501
No log 0.3673 18 1.2809 0.1959 1.2809 1.1318
No log 0.4082 20 1.3741 0.2580 1.3741 1.1722
No log 0.4490 22 1.4955 0.3259 1.4955 1.2229
No log 0.4898 24 1.4972 0.3401 1.4972 1.2236
No log 0.5306 26 1.3235 0.3637 1.3235 1.1505
No log 0.5714 28 1.1132 0.3543 1.1132 1.0551
No log 0.6122 30 1.0751 0.4027 1.0751 1.0369
No log 0.6531 32 1.0345 0.3861 1.0345 1.0171
No log 0.6939 34 1.0100 0.4100 1.0100 1.0050
No log 0.7347 36 1.0503 0.5036 1.0503 1.0249
No log 0.7755 38 1.0785 0.4803 1.0785 1.0385
No log 0.8163 40 1.0774 0.5061 1.0774 1.0380
No log 0.8571 42 1.1462 0.5228 1.1462 1.0706
No log 0.8980 44 1.0185 0.5387 1.0185 1.0092
No log 0.9388 46 0.8627 0.5464 0.8627 0.9288
No log 0.9796 48 0.8831 0.5822 0.8831 0.9397
No log 1.0204 50 0.9451 0.5924 0.9451 0.9721
No log 1.0612 52 1.2021 0.5493 1.2021 1.0964
No log 1.1020 54 1.2551 0.5651 1.2551 1.1203
No log 1.1429 56 1.2425 0.5526 1.2425 1.1147
No log 1.1837 58 1.1139 0.6125 1.1139 1.0554
No log 1.2245 60 1.1369 0.6200 1.1369 1.0662
No log 1.2653 62 1.1350 0.6267 1.1350 1.0654
No log 1.3061 64 0.7768 0.6916 0.7768 0.8813
No log 1.3469 66 0.7567 0.6650 0.7567 0.8699
No log 1.3878 68 0.8937 0.6858 0.8937 0.9454
No log 1.4286 70 1.4626 0.5885 1.4626 1.2094
No log 1.4694 72 1.3264 0.5897 1.3264 1.1517
No log 1.5102 74 0.8758 0.6806 0.8758 0.9359
No log 1.5510 76 0.7118 0.6703 0.7118 0.8437
No log 1.5918 78 0.7563 0.6713 0.7563 0.8696
No log 1.6327 80 0.9554 0.6714 0.9554 0.9774
No log 1.6735 82 1.2364 0.5864 1.2364 1.1119
No log 1.7143 84 1.0234 0.6475 1.0234 1.0116
No log 1.7551 86 0.9160 0.6679 0.9160 0.9571
No log 1.7959 88 0.7473 0.6553 0.7473 0.8645
No log 1.8367 90 0.7465 0.6189 0.7465 0.8640
No log 1.8776 92 0.7544 0.6234 0.7544 0.8686
No log 1.9184 94 0.7585 0.6324 0.7585 0.8709
No log 1.9592 96 0.9091 0.6551 0.9091 0.9535
No log 2.0 98 1.0400 0.6461 1.0400 1.0198
No log 2.0408 100 1.1148 0.6227 1.1148 1.0558
No log 2.0816 102 1.1235 0.6146 1.1235 1.0599
No log 2.1224 104 0.8329 0.6536 0.8329 0.9127
No log 2.1633 106 0.7164 0.6221 0.7164 0.8464
No log 2.2041 108 0.7121 0.6506 0.7121 0.8439
No log 2.2449 110 0.8510 0.6687 0.8510 0.9225
No log 2.2857 112 1.2189 0.5908 1.2189 1.1040
No log 2.3265 114 1.4492 0.5314 1.4492 1.2038
No log 2.3673 116 1.3891 0.5363 1.3891 1.1786
No log 2.4082 118 1.0827 0.6575 1.0827 1.0405
No log 2.4490 120 0.8024 0.7051 0.8024 0.8958
No log 2.4898 122 0.7602 0.7092 0.7602 0.8719
No log 2.5306 124 0.8613 0.7201 0.8613 0.9281
No log 2.5714 126 0.9536 0.7116 0.9536 0.9765
No log 2.6122 128 1.1884 0.6681 1.1884 1.0901
No log 2.6531 130 1.2284 0.6574 1.2284 1.1083
No log 2.6939 132 0.8583 0.7208 0.8583 0.9265
No log 2.7347 134 0.7409 0.7299 0.7409 0.8607
No log 2.7755 136 0.7271 0.7300 0.7271 0.8527
No log 2.8163 138 0.9316 0.6964 0.9316 0.9652
No log 2.8571 140 1.2320 0.6329 1.2320 1.1099
No log 2.8980 142 1.1026 0.6590 1.1026 1.0500
No log 2.9388 144 0.9195 0.6865 0.9195 0.9589
No log 2.9796 146 0.7326 0.6762 0.7326 0.8559
No log 3.0204 148 0.6510 0.6992 0.6510 0.8069
No log 3.0612 150 0.7419 0.6906 0.7419 0.8613
No log 3.1020 152 1.1573 0.6616 1.1573 1.0758
No log 3.1429 154 1.3979 0.6107 1.3979 1.1823
No log 3.1837 156 1.2014 0.6383 1.2014 1.0961
No log 3.2245 158 0.8583 0.6969 0.8583 0.9264
No log 3.2653 160 0.6302 0.7346 0.6302 0.7938
No log 3.3061 162 0.5912 0.7201 0.5912 0.7689
No log 3.3469 164 0.6296 0.7198 0.6296 0.7935
No log 3.3878 166 0.8100 0.7051 0.8100 0.9000
No log 3.4286 168 0.8298 0.7051 0.8298 0.9109
No log 3.4694 170 0.8555 0.7033 0.8555 0.9249
No log 3.5102 172 0.7326 0.7244 0.7326 0.8559
No log 3.5510 174 0.6320 0.7309 0.6320 0.7950
No log 3.5918 176 0.6803 0.7186 0.6803 0.8248
No log 3.6327 178 0.8196 0.6981 0.8196 0.9053
No log 3.6735 180 0.9180 0.6928 0.9180 0.9581
No log 3.7143 182 0.8003 0.7037 0.8003 0.8946
No log 3.7551 184 0.6613 0.7179 0.6613 0.8132
No log 3.7959 186 0.6708 0.7166 0.6708 0.8190
No log 3.8367 188 0.8253 0.7081 0.8253 0.9085
No log 3.8776 190 0.8607 0.6982 0.8607 0.9278
No log 3.9184 192 0.7733 0.7159 0.7733 0.8794
No log 3.9592 194 0.6324 0.7146 0.6324 0.7952
No log 4.0 196 0.6061 0.7153 0.6061 0.7785
No log 4.0408 198 0.6336 0.7057 0.6336 0.7960
No log 4.0816 200 0.7870 0.6987 0.7870 0.8871
No log 4.1224 202 1.0684 0.6318 1.0684 1.0337
No log 4.1633 204 1.1481 0.6385 1.1481 1.0715
No log 4.2041 206 1.0294 0.6472 1.0294 1.0146
No log 4.2449 208 0.8496 0.6779 0.8496 0.9217
No log 4.2857 210 0.7545 0.6980 0.7545 0.8686
No log 4.3265 212 0.7360 0.6815 0.7360 0.8579
No log 4.3673 214 0.7255 0.6891 0.7255 0.8518
No log 4.4082 216 0.7662 0.7006 0.7662 0.8754
No log 4.4490 218 0.8519 0.7011 0.8519 0.9230
No log 4.4898 220 0.8570 0.6662 0.8570 0.9257
No log 4.5306 222 0.7876 0.7100 0.7876 0.8875
No log 4.5714 224 0.7797 0.7176 0.7797 0.8830
No log 4.6122 226 0.7378 0.7275 0.7378 0.8590
No log 4.6531 228 0.7957 0.7136 0.7957 0.8920
No log 4.6939 230 0.9391 0.6873 0.9391 0.9691
No log 4.7347 232 0.9259 0.6721 0.9259 0.9622
No log 4.7755 234 0.8592 0.6988 0.8592 0.9269
No log 4.8163 236 0.7106 0.7132 0.7106 0.8430
No log 4.8571 238 0.6289 0.7317 0.6289 0.7930
No log 4.8980 240 0.6071 0.7369 0.6071 0.7791
No log 4.9388 242 0.6351 0.7496 0.6351 0.7969
No log 4.9796 244 0.7338 0.7233 0.7338 0.8566
No log 5.0204 246 0.7491 0.7240 0.7491 0.8655
No log 5.0612 248 0.7499 0.7206 0.7499 0.8660
No log 5.1020 250 0.7716 0.7075 0.7716 0.8784
No log 5.1429 252 0.8157 0.7015 0.8157 0.9031
No log 5.1837 254 0.8175 0.6850 0.8175 0.9042
No log 5.2245 256 0.8737 0.6942 0.8737 0.9347
No log 5.2653 258 0.8731 0.6778 0.8731 0.9344
No log 5.3061 260 0.7754 0.7180 0.7754 0.8806
No log 5.3469 262 0.7436 0.7120 0.7436 0.8623
No log 5.3878 264 0.7625 0.7186 0.7625 0.8732
No log 5.4286 266 0.8260 0.6791 0.8260 0.9088
No log 5.4694 268 0.9986 0.6741 0.9986 0.9993
No log 5.5102 270 1.0601 0.6660 1.0601 1.0296
No log 5.5510 272 0.9313 0.6779 0.9313 0.9650
No log 5.5918 274 0.7793 0.7089 0.7793 0.8828
No log 5.6327 276 0.7401 0.7419 0.7401 0.8603
No log 5.6735 278 0.6528 0.7331 0.6528 0.8080
No log 5.7143 280 0.6000 0.7574 0.6000 0.7746
No log 5.7551 282 0.5939 0.7330 0.5939 0.7706
No log 5.7959 284 0.6258 0.7301 0.6258 0.7911
No log 5.8367 286 0.7113 0.7252 0.7113 0.8434
No log 5.8776 288 0.7162 0.7194 0.7162 0.8463
No log 5.9184 290 0.6818 0.6969 0.6818 0.8257
No log 5.9592 292 0.6694 0.7017 0.6694 0.8182
No log 6.0 294 0.6972 0.7067 0.6972 0.8350
No log 6.0408 296 0.6883 0.7206 0.6883 0.8296
No log 6.0816 298 0.6451 0.7345 0.6451 0.8032
No log 6.1224 300 0.6594 0.7529 0.6594 0.8120
No log 6.1633 302 0.7431 0.7321 0.7431 0.8620
No log 6.2041 304 0.8393 0.6991 0.8393 0.9162
No log 6.2449 306 0.9125 0.6779 0.9125 0.9553
No log 6.2857 308 0.8677 0.6901 0.8677 0.9315
No log 6.3265 310 0.7969 0.7071 0.7969 0.8927
No log 6.3673 312 0.8039 0.6898 0.8039 0.8966
No log 6.4082 314 0.8589 0.6914 0.8589 0.9267
No log 6.4490 316 0.8248 0.6832 0.8248 0.9082
No log 6.4898 318 0.7762 0.6759 0.7762 0.8810
No log 6.5306 320 0.7947 0.6784 0.7947 0.8915
No log 6.5714 322 0.8493 0.6886 0.8493 0.9216
No log 6.6122 324 0.9227 0.6781 0.9227 0.9605
No log 6.6531 326 0.9214 0.6711 0.9214 0.9599
No log 6.6939 328 0.8293 0.7289 0.8293 0.9106
No log 6.7347 330 0.7093 0.7155 0.7093 0.8422
No log 6.7755 332 0.6562 0.7315 0.6562 0.8101
No log 6.8163 334 0.6496 0.7249 0.6496 0.8060
No log 6.8571 336 0.6710 0.7249 0.6710 0.8191
No log 6.8980 338 0.7083 0.7160 0.7083 0.8416
No log 6.9388 340 0.7530 0.6958 0.7530 0.8678
No log 6.9796 342 0.7684 0.7134 0.7684 0.8766
No log 7.0204 344 0.7789 0.7214 0.7789 0.8825
No log 7.0612 346 0.7298 0.7098 0.7298 0.8543
No log 7.1020 348 0.6712 0.7161 0.6712 0.8193
No log 7.1429 350 0.6219 0.7426 0.6219 0.7886
No log 7.1837 352 0.5913 0.7398 0.5913 0.7690
No log 7.2245 354 0.5905 0.7416 0.5905 0.7684
No log 7.2653 356 0.6120 0.7406 0.6120 0.7823
No log 7.3061 358 0.6742 0.7243 0.6742 0.8211
No log 7.3469 360 0.7798 0.7196 0.7798 0.8831
No log 7.3878 362 0.8374 0.7150 0.8374 0.9151
No log 7.4286 364 0.8411 0.7096 0.8411 0.9171
No log 7.4694 366 0.7918 0.7132 0.7918 0.8898
No log 7.5102 368 0.7067 0.7194 0.7067 0.8406
No log 7.5510 370 0.6406 0.7216 0.6406 0.8004
No log 7.5918 372 0.6013 0.7219 0.6013 0.7755
No log 7.6327 374 0.5907 0.7217 0.5907 0.7686
No log 7.6735 376 0.5955 0.7217 0.5955 0.7717
No log 7.7143 378 0.6179 0.7152 0.6179 0.7861
No log 7.7551 380 0.6671 0.7082 0.6671 0.8168
No log 7.7959 382 0.7224 0.6990 0.7224 0.8499
No log 7.8367 384 0.8196 0.6892 0.8196 0.9053
No log 7.8776 386 0.9037 0.6907 0.9037 0.9506
No log 7.9184 388 0.9363 0.6791 0.9363 0.9676
No log 7.9592 390 0.9154 0.7037 0.9154 0.9568
No log 8.0 392 0.8541 0.6907 0.8541 0.9241
No log 8.0408 394 0.7727 0.7003 0.7727 0.8790
No log 8.0816 396 0.6938 0.7070 0.6938 0.8330
No log 8.1224 398 0.6581 0.7326 0.6581 0.8112
No log 8.1633 400 0.6544 0.7241 0.6544 0.8090
No log 8.2041 402 0.6792 0.7341 0.6792 0.8241
No log 8.2449 404 0.7305 0.7026 0.7305 0.8547
No log 8.2857 406 0.7951 0.6936 0.7951 0.8917
No log 8.3265 408 0.8372 0.6723 0.8372 0.9150
No log 8.3673 410 0.8661 0.6731 0.8661 0.9307
No log 8.4082 412 0.8768 0.6746 0.8768 0.9364
No log 8.4490 414 0.8597 0.6731 0.8597 0.9272
No log 8.4898 416 0.8354 0.6898 0.8354 0.9140
No log 8.5306 418 0.8227 0.6898 0.8227 0.9070
No log 8.5714 420 0.8131 0.6831 0.8131 0.9017
No log 8.6122 422 0.7934 0.6875 0.7934 0.8907
No log 8.6531 424 0.7978 0.6875 0.7978 0.8932
No log 8.6939 426 0.8107 0.6875 0.8107 0.9004
No log 8.7347 428 0.8216 0.6941 0.8216 0.9064
No log 8.7755 430 0.8243 0.6887 0.8243 0.9079
No log 8.8163 432 0.8103 0.7011 0.8103 0.9001
No log 8.8571 434 0.7980 0.7056 0.7980 0.8933
No log 8.8980 436 0.7698 0.7117 0.7698 0.8774
No log 8.9388 438 0.7371 0.7050 0.7371 0.8585
No log 8.9796 440 0.7182 0.7244 0.7182 0.8475
No log 9.0204 442 0.7139 0.7162 0.7139 0.8449
No log 9.0612 444 0.7242 0.7132 0.7242 0.8510
No log 9.1020 446 0.7358 0.7070 0.7358 0.8578
No log 9.1429 448 0.7508 0.7050 0.7508 0.8665
No log 9.1837 450 0.7691 0.7030 0.7691 0.8770
No log 9.2245 452 0.7851 0.7037 0.7851 0.8861
No log 9.2653 454 0.7984 0.7196 0.7984 0.8935
No log 9.3061 456 0.8057 0.7196 0.8057 0.8976
No log 9.3469 458 0.8082 0.7196 0.8082 0.8990
No log 9.3878 460 0.8121 0.7136 0.8121 0.9011
No log 9.4286 462 0.8178 0.7136 0.8178 0.9043
No log 9.4694 464 0.8205 0.6974 0.8205 0.9058
No log 9.5102 466 0.8139 0.7015 0.8139 0.9022
No log 9.5510 468 0.8012 0.6976 0.8012 0.8951
No log 9.5918 470 0.7964 0.6967 0.7964 0.8924
No log 9.6327 472 0.7953 0.6924 0.7953 0.8918
No log 9.6735 474 0.7941 0.6924 0.7941 0.8911
No log 9.7143 476 0.7907 0.6924 0.7907 0.8892
No log 9.7551 478 0.7881 0.6924 0.7881 0.8878
No log 9.7959 480 0.7879 0.6924 0.7879 0.8876
No log 9.8367 482 0.7871 0.6924 0.7871 0.8872
No log 9.8776 484 0.7855 0.6924 0.7855 0.8863
No log 9.9184 486 0.7848 0.6924 0.7848 0.8859
No log 9.9592 488 0.7843 0.6924 0.7843 0.8856
No log 10.0 490 0.7846 0.6924 0.7846 0.8858

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits6_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k12_task5_organization

Finetuned
(4023)
this model