ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k17_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7581
  • Qwk: 0.2784
  • Mse: 0.7581
  • Rmse: 0.8707

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0233 2 2.5766 -0.0262 2.5766 1.6052
No log 0.0465 4 1.3048 0.0997 1.3048 1.1423
No log 0.0698 6 0.9277 -0.0660 0.9277 0.9632
No log 0.0930 8 0.8725 0.0637 0.8725 0.9341
No log 0.1163 10 0.8817 0.2589 0.8817 0.9390
No log 0.1395 12 0.8682 0.2206 0.8682 0.9318
No log 0.1628 14 0.9059 0.1775 0.9059 0.9518
No log 0.1860 16 0.8248 0.2336 0.8248 0.9082
No log 0.2093 18 0.8325 0.2405 0.8325 0.9124
No log 0.2326 20 0.7714 0.2405 0.7714 0.8783
No log 0.2558 22 0.7829 0.2405 0.7829 0.8848
No log 0.2791 24 0.7877 0.1139 0.7877 0.8875
No log 0.3023 26 0.7470 0.2024 0.7470 0.8643
No log 0.3256 28 0.7358 0.2713 0.7358 0.8578
No log 0.3488 30 0.7811 0.0764 0.7811 0.8838
No log 0.3721 32 0.7202 0.2540 0.7202 0.8486
No log 0.3953 34 0.6927 0.3081 0.6927 0.8323
No log 0.4186 36 0.7165 0.2652 0.7165 0.8465
No log 0.4419 38 0.6843 0.2783 0.6843 0.8272
No log 0.4651 40 0.7514 0.2545 0.7514 0.8668
No log 0.4884 42 0.9163 0.1973 0.9163 0.9572
No log 0.5116 44 1.1367 0.0452 1.1367 1.0662
No log 0.5349 46 1.2227 0.0479 1.2227 1.1057
No log 0.5581 48 1.2830 0.0479 1.2830 1.1327
No log 0.5814 50 1.0309 0.2263 1.0309 1.0153
No log 0.6047 52 0.7989 0.1372 0.7989 0.8938
No log 0.6279 54 0.7324 0.0810 0.7324 0.8558
No log 0.6512 56 0.7059 0.1498 0.7059 0.8402
No log 0.6744 58 0.7048 0.2345 0.7048 0.8396
No log 0.6977 60 0.7053 0.1962 0.7053 0.8398
No log 0.7209 62 0.7357 0.2353 0.7357 0.8577
No log 0.7442 64 0.7822 0.1647 0.7822 0.8844
No log 0.7674 66 0.8044 0.1977 0.8044 0.8969
No log 0.7907 68 0.9603 0.2008 0.9603 0.9799
No log 0.8140 70 1.1174 0.0989 1.1174 1.0571
No log 0.8372 72 1.1690 0.0731 1.1690 1.0812
No log 0.8605 74 1.2111 0.0993 1.2111 1.1005
No log 0.8837 76 1.0874 0.2113 1.0874 1.0428
No log 0.9070 78 0.8675 0.2402 0.8675 0.9314
No log 0.9302 80 0.7493 0.1221 0.7493 0.8656
No log 0.9535 82 0.7108 0.1456 0.7108 0.8431
No log 0.9767 84 0.7257 0.1359 0.7257 0.8519
No log 1.0 86 0.7426 0.0295 0.7426 0.8617
No log 1.0233 88 0.8246 0.0026 0.8246 0.9081
No log 1.0465 90 0.9476 0.1700 0.9476 0.9734
No log 1.0698 92 0.9304 0.1649 0.9304 0.9646
No log 1.0930 94 0.9152 0.1290 0.9152 0.9567
No log 1.1163 96 0.9543 0.1281 0.9543 0.9769
No log 1.1395 98 0.9700 0.1281 0.9700 0.9849
No log 1.1628 100 0.9354 0.0501 0.9354 0.9671
No log 1.1860 102 0.9515 0.0909 0.9515 0.9754
No log 1.2093 104 0.9541 0.1352 0.9541 0.9768
No log 1.2326 106 0.9029 0.0 0.9029 0.9502
No log 1.2558 108 0.8932 0.0344 0.8932 0.9451
No log 1.2791 110 0.9177 -0.0025 0.9177 0.9580
No log 1.3023 112 1.0074 0.0957 1.0074 1.0037
No log 1.3256 114 1.2305 -0.0423 1.2305 1.1093
No log 1.3488 116 1.2091 -0.0423 1.2091 1.0996
No log 1.3721 118 1.1576 0.0092 1.1576 1.0759
No log 1.3953 120 1.1559 0.0356 1.1559 1.0751
No log 1.4186 122 1.2089 0.0401 1.2089 1.0995
No log 1.4419 124 1.2125 0.0666 1.2125 1.1011
No log 1.4651 126 1.1850 0.0412 1.1850 1.0886
No log 1.4884 128 1.2220 0.0461 1.2220 1.1054
No log 1.5116 130 1.0626 0.1220 1.0626 1.0308
No log 1.5349 132 0.9969 0.1228 0.9969 0.9984
No log 1.5581 134 0.9898 0.0906 0.9898 0.9949
No log 1.5814 136 1.1055 0.0915 1.1055 1.0514
No log 1.6047 138 1.3702 -0.0168 1.3702 1.1706
No log 1.6279 140 1.5379 0.0566 1.5379 1.2401
No log 1.6512 142 1.4445 0.0566 1.4445 1.2019
No log 1.6744 144 1.2036 0.0745 1.2036 1.0971
No log 1.6977 146 0.9710 0.1234 0.9710 0.9854
No log 1.7209 148 0.8952 0.1259 0.8952 0.9462
No log 1.7442 150 0.8808 0.0861 0.8808 0.9385
No log 1.7674 152 0.9476 0.0955 0.9476 0.9735
No log 1.7907 154 0.9940 0.0964 0.9940 0.9970
No log 1.8140 156 0.9265 0.1310 0.9265 0.9626
No log 1.8372 158 0.8563 -0.0443 0.8563 0.9253
No log 1.8605 160 0.8462 -0.0526 0.8462 0.9199
No log 1.8837 162 0.8589 0.0827 0.8589 0.9268
No log 1.9070 164 1.0307 0.1269 1.0307 1.0152
No log 1.9302 166 1.1311 0.1277 1.1311 1.0635
No log 1.9535 168 1.0214 0.1228 1.0214 1.0107
No log 1.9767 170 0.9145 0.1052 0.9145 0.9563
No log 2.0 172 0.8679 0.2563 0.8679 0.9316
No log 2.0233 174 0.8764 0.1090 0.8764 0.9361
No log 2.0465 176 0.9630 0.0860 0.9630 0.9813
No log 2.0698 178 0.9598 0.1201 0.9598 0.9797
No log 2.0930 180 0.8869 0.1586 0.8869 0.9418
No log 2.1163 182 0.8424 0.0301 0.8424 0.9178
No log 2.1395 184 0.8335 0.1935 0.8335 0.9130
No log 2.1628 186 0.8496 0.1786 0.8496 0.9217
No log 2.1860 188 0.8448 0.1487 0.8448 0.9191
No log 2.2093 190 0.8507 0.1988 0.8507 0.9223
No log 2.2326 192 0.8591 0.1786 0.8591 0.9269
No log 2.2558 194 0.8694 0.2149 0.8694 0.9324
No log 2.2791 196 0.8457 -0.0511 0.8457 0.9196
No log 2.3023 198 0.8516 -0.0479 0.8516 0.9228
No log 2.3256 200 0.8203 0.1746 0.8203 0.9057
No log 2.3488 202 0.8090 0.2078 0.8090 0.8995
No log 2.3721 204 0.8163 0.2092 0.8163 0.9035
No log 2.3953 206 0.8092 0.2092 0.8092 0.8995
No log 2.4186 208 0.7861 0.2170 0.7861 0.8866
No log 2.4419 210 0.7891 0.0778 0.7891 0.8883
No log 2.4651 212 0.7674 0.1176 0.7674 0.8760
No log 2.4884 214 0.7625 0.0778 0.7625 0.8732
No log 2.5116 216 0.8281 0.2214 0.8281 0.9100
No log 2.5349 218 0.9561 0.1798 0.9561 0.9778
No log 2.5581 220 0.9271 0.1528 0.9271 0.9629
No log 2.5814 222 0.7981 0.2019 0.7981 0.8934
No log 2.6047 224 0.7691 0.2505 0.7691 0.8770
No log 2.6279 226 0.7465 0.2838 0.7465 0.8640
No log 2.6512 228 0.7976 0.3367 0.7976 0.8931
No log 2.6744 230 0.8891 0.2091 0.8891 0.9429
No log 2.6977 232 0.8255 0.2401 0.8255 0.9086
No log 2.7209 234 0.7037 0.3229 0.7037 0.8389
No log 2.7442 236 0.7221 0.3665 0.7221 0.8498
No log 2.7674 238 0.7472 0.3891 0.7472 0.8644
No log 2.7907 240 0.7024 0.3478 0.7024 0.8381
No log 2.8140 242 0.7413 0.3829 0.7413 0.8610
No log 2.8372 244 0.8110 0.2476 0.8110 0.9006
No log 2.8605 246 0.7626 0.3367 0.7626 0.8733
No log 2.8837 248 0.7485 0.3946 0.7485 0.8652
No log 2.9070 250 0.7346 0.3762 0.7346 0.8571
No log 2.9302 252 0.7611 0.4461 0.7611 0.8724
No log 2.9535 254 0.9431 0.2729 0.9431 0.9711
No log 2.9767 256 1.0029 0.1960 1.0029 1.0014
No log 3.0 258 0.8975 0.2267 0.8975 0.9474
No log 3.0233 260 0.7745 0.3215 0.7745 0.8801
No log 3.0465 262 0.6845 0.4128 0.6845 0.8273
No log 3.0698 264 0.6832 0.4700 0.6832 0.8266
No log 3.0930 266 0.6825 0.5021 0.6825 0.8262
No log 3.1163 268 0.7292 0.3167 0.7292 0.8539
No log 3.1395 270 0.8328 0.2554 0.8328 0.9126
No log 3.1628 272 0.7442 0.3937 0.7442 0.8627
No log 3.1860 274 0.6633 0.5003 0.6633 0.8145
No log 3.2093 276 0.7495 0.4294 0.7495 0.8657
No log 3.2326 278 0.7632 0.3723 0.7632 0.8736
No log 3.2558 280 0.6939 0.3891 0.6939 0.8330
No log 3.2791 282 0.6780 0.4299 0.6780 0.8234
No log 3.3023 284 0.8335 0.2910 0.8335 0.9130
No log 3.3256 286 0.9320 0.2310 0.9320 0.9654
No log 3.3488 288 0.8999 0.2044 0.8999 0.9486
No log 3.3721 290 0.7865 0.2793 0.7865 0.8868
No log 3.3953 292 0.7007 0.3213 0.7007 0.8371
No log 3.4186 294 0.6988 0.2973 0.6988 0.8360
No log 3.4419 296 0.6972 0.2884 0.6972 0.8350
No log 3.4651 298 0.7222 0.3050 0.7222 0.8498
No log 3.4884 300 0.7775 0.2581 0.7775 0.8818
No log 3.5116 302 0.7944 0.2133 0.7944 0.8913
No log 3.5349 304 0.7811 0.3051 0.7811 0.8838
No log 3.5581 306 0.6754 0.3089 0.6754 0.8218
No log 3.5814 308 0.6638 0.4705 0.6638 0.8147
No log 3.6047 310 0.6778 0.4158 0.6778 0.8233
No log 3.6279 312 0.6472 0.4637 0.6472 0.8045
No log 3.6512 314 0.6410 0.4229 0.6410 0.8006
No log 3.6744 316 0.6464 0.4229 0.6464 0.8040
No log 3.6977 318 0.6530 0.4482 0.6530 0.8081
No log 3.7209 320 0.6602 0.3970 0.6602 0.8125
No log 3.7442 322 0.6620 0.4660 0.6620 0.8136
No log 3.7674 324 0.6661 0.4576 0.6661 0.8161
No log 3.7907 326 0.6749 0.4576 0.6749 0.8216
No log 3.8140 328 0.6560 0.4837 0.6560 0.8099
No log 3.8372 330 0.6522 0.4240 0.6522 0.8076
No log 3.8605 332 0.6648 0.4527 0.6648 0.8154
No log 3.8837 334 0.7872 0.2830 0.7872 0.8873
No log 3.9070 336 0.8976 0.2555 0.8976 0.9474
No log 3.9302 338 0.7960 0.3052 0.7960 0.8922
No log 3.9535 340 0.7099 0.3448 0.7099 0.8425
No log 3.9767 342 0.7054 0.3448 0.7054 0.8399
No log 4.0 344 0.8147 0.2554 0.8147 0.9026
No log 4.0233 346 0.8069 0.2554 0.8069 0.8983
No log 4.0465 348 0.6787 0.4411 0.6787 0.8238
No log 4.0698 350 0.6373 0.4249 0.6373 0.7983
No log 4.0930 352 0.6580 0.4193 0.6580 0.8112
No log 4.1163 354 0.8072 0.3290 0.8072 0.8984
No log 4.1395 356 1.0929 0.2471 1.0929 1.0454
No log 4.1628 358 1.1765 0.2395 1.1765 1.0847
No log 4.1860 360 1.2651 0.2120 1.2651 1.1248
No log 4.2093 362 1.0444 0.3135 1.0444 1.0220
No log 4.2326 364 0.9815 0.3486 0.9815 0.9907
No log 4.2558 366 1.0060 0.3279 1.0060 1.0030
No log 4.2791 368 0.8607 0.3767 0.8607 0.9277
No log 4.3023 370 0.7960 0.3727 0.7960 0.8922
No log 4.3256 372 0.7415 0.3284 0.7415 0.8611
No log 4.3488 374 0.7273 0.3186 0.7273 0.8528
No log 4.3721 376 0.7488 0.3561 0.7488 0.8653
No log 4.3953 378 0.7345 0.3607 0.7345 0.8571
No log 4.4186 380 0.7290 0.3467 0.7290 0.8538
No log 4.4419 382 0.7282 0.3433 0.7282 0.8533
No log 4.4651 384 0.7597 0.3512 0.7597 0.8716
No log 4.4884 386 0.8887 0.3398 0.8887 0.9427
No log 4.5116 388 0.9348 0.3135 0.9348 0.9668
No log 4.5349 390 0.8767 0.3165 0.8767 0.9363
No log 4.5581 392 0.7472 0.3373 0.7472 0.8644
No log 4.5814 394 0.7162 0.3935 0.7162 0.8463
No log 4.6047 396 0.7176 0.3590 0.7176 0.8471
No log 4.6279 398 0.7517 0.3410 0.7517 0.8670
No log 4.6512 400 0.7965 0.3576 0.7965 0.8925
No log 4.6744 402 0.7653 0.3373 0.7653 0.8748
No log 4.6977 404 0.7162 0.4186 0.7162 0.8463
No log 4.7209 406 0.7582 0.4081 0.7582 0.8707
No log 4.7442 408 0.7394 0.3793 0.7394 0.8599
No log 4.7674 410 0.7146 0.4182 0.7146 0.8454
No log 4.7907 412 0.7402 0.3457 0.7402 0.8604
No log 4.8140 414 0.7613 0.3264 0.7613 0.8725
No log 4.8372 416 0.7281 0.3178 0.7281 0.8533
No log 4.8605 418 0.6701 0.4475 0.6701 0.8186
No log 4.8837 420 0.6703 0.4147 0.6703 0.8187
No log 4.9070 422 0.6862 0.4434 0.6862 0.8284
No log 4.9302 424 0.6971 0.4125 0.6971 0.8349
No log 4.9535 426 0.6936 0.4620 0.6936 0.8328
No log 4.9767 428 0.7142 0.4179 0.7142 0.8451
No log 5.0 430 0.7358 0.4292 0.7358 0.8578
No log 5.0233 432 0.7597 0.3095 0.7597 0.8716
No log 5.0465 434 0.7206 0.4314 0.7206 0.8489
No log 5.0698 436 0.6856 0.4137 0.6856 0.8280
No log 5.0930 438 0.7075 0.2971 0.7075 0.8411
No log 5.1163 440 0.7217 0.2624 0.7217 0.8495
No log 5.1395 442 0.7222 0.4137 0.7222 0.8498
No log 5.1628 444 0.7396 0.4342 0.7396 0.8600
No log 5.1860 446 0.7391 0.4342 0.7391 0.8597
No log 5.2093 448 0.7486 0.3153 0.7486 0.8652
No log 5.2326 450 0.7543 0.2819 0.7543 0.8685
No log 5.2558 452 0.7314 0.3835 0.7314 0.8552
No log 5.2791 454 0.7410 0.4828 0.7410 0.8608
No log 5.3023 456 0.7291 0.4828 0.7291 0.8539
No log 5.3256 458 0.7140 0.3293 0.7140 0.8450
No log 5.3488 460 0.7418 0.2680 0.7418 0.8613
No log 5.3721 462 0.7968 0.2652 0.7968 0.8926
No log 5.3953 464 0.8161 0.2651 0.8161 0.9034
No log 5.4186 466 0.7548 0.2571 0.7548 0.8688
No log 5.4419 468 0.7198 0.3810 0.7198 0.8484
No log 5.4651 470 0.7237 0.4244 0.7237 0.8507
No log 5.4884 472 0.7268 0.3347 0.7268 0.8525
No log 5.5116 474 0.7793 0.2870 0.7793 0.8828
No log 5.5349 476 0.8577 0.2576 0.8577 0.9261
No log 5.5581 478 0.8467 0.2576 0.8467 0.9201
No log 5.5814 480 0.7569 0.2624 0.7569 0.8700
No log 5.6047 482 0.7430 0.3448 0.7430 0.8620
No log 5.6279 484 0.7528 0.3274 0.7528 0.8676
No log 5.6512 486 0.7245 0.3910 0.7245 0.8512
No log 5.6744 488 0.7334 0.2281 0.7334 0.8564
No log 5.6977 490 0.7526 0.3202 0.7526 0.8675
No log 5.7209 492 0.7664 0.3410 0.7664 0.8754
No log 5.7442 494 0.7822 0.3114 0.7822 0.8844
No log 5.7674 496 0.7429 0.2268 0.7429 0.8619
No log 5.7907 498 0.7164 0.4287 0.7164 0.8464
0.3619 5.8140 500 0.7243 0.4006 0.7243 0.8511
0.3619 5.8372 502 0.7255 0.3738 0.7255 0.8518
0.3619 5.8605 504 0.7168 0.3883 0.7168 0.8466
0.3619 5.8837 506 0.7309 0.3252 0.7309 0.8550
0.3619 5.9070 508 0.7655 0.2890 0.7655 0.8749
0.3619 5.9302 510 0.7581 0.2784 0.7581 0.8707

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run2_AugV5_k17_task7_organization

Finetuned
(4019)
this model