ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k17_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7284
  • Qwk: 0.6739
  • Mse: 0.7284
  • Rmse: 0.8534

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0241 2 5.2338 -0.0020 5.2338 2.2878
No log 0.0482 4 3.3228 0.0669 3.3228 1.8229
No log 0.0723 6 2.0603 0.0830 2.0603 1.4354
No log 0.0964 8 1.5058 0.1544 1.5058 1.2271
No log 0.1205 10 1.2428 0.2021 1.2428 1.1148
No log 0.1446 12 1.2197 0.2003 1.2197 1.1044
No log 0.1687 14 1.2608 0.1656 1.2608 1.1229
No log 0.1928 16 1.2815 0.1517 1.2815 1.1320
No log 0.2169 18 1.3516 0.1248 1.3516 1.1626
No log 0.2410 20 1.5152 0.0901 1.5152 1.2309
No log 0.2651 22 1.9306 0.0964 1.9306 1.3894
No log 0.2892 24 1.8734 0.0954 1.8734 1.3687
No log 0.3133 26 1.3864 0.2129 1.3864 1.1775
No log 0.3373 28 1.0915 0.3515 1.0915 1.0447
No log 0.3614 30 1.2243 0.2483 1.2243 1.1065
No log 0.3855 32 1.2387 0.2688 1.2387 1.1130
No log 0.4096 34 1.1703 0.2845 1.1703 1.0818
No log 0.4337 36 1.2476 0.3007 1.2476 1.1170
No log 0.4578 38 1.7231 0.2059 1.7231 1.3127
No log 0.4819 40 1.8858 0.1991 1.8858 1.3732
No log 0.5060 42 1.6360 0.2655 1.6360 1.2790
No log 0.5301 44 1.3310 0.3349 1.3310 1.1537
No log 0.5542 46 1.4084 0.3266 1.4084 1.1868
No log 0.5783 48 1.6411 0.2300 1.6411 1.2811
No log 0.6024 50 1.4233 0.2810 1.4233 1.1930
No log 0.6265 52 1.1994 0.3126 1.1994 1.0952
No log 0.6506 54 1.1703 0.3148 1.1703 1.0818
No log 0.6747 56 1.1909 0.3350 1.1909 1.0913
No log 0.6988 58 1.3134 0.2567 1.3134 1.1460
No log 0.7229 60 1.3530 0.2282 1.3530 1.1632
No log 0.7470 62 1.3115 0.2866 1.3115 1.1452
No log 0.7711 64 1.2840 0.2910 1.2840 1.1331
No log 0.7952 66 1.3600 0.2683 1.3600 1.1662
No log 0.8193 68 1.4143 0.2489 1.4143 1.1892
No log 0.8434 70 1.4996 0.1988 1.4996 1.2246
No log 0.8675 72 1.3785 0.2380 1.3785 1.1741
No log 0.8916 74 1.2849 0.3069 1.2849 1.1335
No log 0.9157 76 1.2934 0.3031 1.2934 1.1373
No log 0.9398 78 1.3283 0.2829 1.3283 1.1525
No log 0.9639 80 1.3259 0.2609 1.3259 1.1515
No log 0.9880 82 1.3873 0.2342 1.3873 1.1778
No log 1.0120 84 1.4126 0.1726 1.4126 1.1885
No log 1.0361 86 1.7363 0.1527 1.7363 1.3177
No log 1.0602 88 2.1388 0.0655 2.1388 1.4625
No log 1.0843 90 1.9523 0.1201 1.9523 1.3972
No log 1.1084 92 1.7822 0.1204 1.7822 1.3350
No log 1.1325 94 1.4520 0.2462 1.4520 1.2050
No log 1.1566 96 1.0825 0.3790 1.0825 1.0405
No log 1.1807 98 0.8900 0.4624 0.8900 0.9434
No log 1.2048 100 0.8534 0.5522 0.8534 0.9238
No log 1.2289 102 0.8411 0.6034 0.8411 0.9171
No log 1.2530 104 1.1292 0.5396 1.1292 1.0626
No log 1.2771 106 1.2921 0.5036 1.2921 1.1367
No log 1.3012 108 1.0655 0.5578 1.0655 1.0323
No log 1.3253 110 0.9484 0.5903 0.9484 0.9738
No log 1.3494 112 0.8718 0.5928 0.8718 0.9337
No log 1.3735 114 0.7670 0.6716 0.7670 0.8758
No log 1.3976 116 0.7376 0.6780 0.7376 0.8588
No log 1.4217 118 0.7264 0.6286 0.7264 0.8523
No log 1.4458 120 0.7359 0.6101 0.7359 0.8579
No log 1.4699 122 0.7530 0.5919 0.7530 0.8677
No log 1.4940 124 0.7843 0.5915 0.7843 0.8856
No log 1.5181 126 0.8107 0.5797 0.8107 0.9004
No log 1.5422 128 0.8858 0.5171 0.8858 0.9412
No log 1.5663 130 0.9032 0.5130 0.9032 0.9503
No log 1.5904 132 0.9092 0.4985 0.9092 0.9535
No log 1.6145 134 0.8654 0.5875 0.8654 0.9303
No log 1.6386 136 0.8041 0.6334 0.8041 0.8967
No log 1.6627 138 0.8396 0.5598 0.8396 0.9163
No log 1.6867 140 0.8690 0.4940 0.8690 0.9322
No log 1.7108 142 0.8078 0.5669 0.8078 0.8988
No log 1.7349 144 0.8224 0.5450 0.8224 0.9069
No log 1.7590 146 0.9448 0.4700 0.9448 0.9720
No log 1.7831 148 1.1091 0.4121 1.1091 1.0531
No log 1.8072 150 1.0461 0.4599 1.0461 1.0228
No log 1.8313 152 0.8314 0.5317 0.8314 0.9118
No log 1.8554 154 0.7456 0.5735 0.7456 0.8635
No log 1.8795 156 0.7378 0.5780 0.7378 0.8589
No log 1.9036 158 0.7447 0.6146 0.7447 0.8630
No log 1.9277 160 0.7306 0.6396 0.7306 0.8548
No log 1.9518 162 0.7658 0.6376 0.7658 0.8751
No log 1.9759 164 0.7733 0.6391 0.7733 0.8794
No log 2.0 166 0.8052 0.6172 0.8052 0.8973
No log 2.0241 168 0.8829 0.5722 0.8829 0.9396
No log 2.0482 170 0.8179 0.5865 0.8179 0.9044
No log 2.0723 172 0.7707 0.5957 0.7707 0.8779
No log 2.0964 174 0.7459 0.6005 0.7459 0.8637
No log 2.1205 176 0.7611 0.6022 0.7611 0.8724
No log 2.1446 178 0.7775 0.6039 0.7775 0.8817
No log 2.1687 180 0.7809 0.6016 0.7809 0.8837
No log 2.1928 182 0.7745 0.6401 0.7745 0.8800
No log 2.2169 184 0.8037 0.6128 0.8037 0.8965
No log 2.2410 186 0.9955 0.5118 0.9955 0.9978
No log 2.2651 188 1.1210 0.3948 1.1210 1.0588
No log 2.2892 190 1.0904 0.3985 1.0904 1.0442
No log 2.3133 192 1.0097 0.4250 1.0097 1.0049
No log 2.3373 194 1.1075 0.4307 1.1075 1.0524
No log 2.3614 196 1.0301 0.4723 1.0301 1.0149
No log 2.3855 198 0.8574 0.5265 0.8574 0.9260
No log 2.4096 200 0.7640 0.5614 0.7640 0.8741
No log 2.4337 202 0.9202 0.5288 0.9202 0.9593
No log 2.4578 204 1.2263 0.5035 1.2263 1.1074
No log 2.4819 206 1.2543 0.5040 1.2543 1.1199
No log 2.5060 208 0.9587 0.5836 0.9587 0.9792
No log 2.5301 210 0.7779 0.6287 0.7779 0.8820
No log 2.5542 212 0.7427 0.5944 0.7427 0.8618
No log 2.5783 214 0.7632 0.6377 0.7632 0.8736
No log 2.6024 216 0.7477 0.6006 0.7477 0.8647
No log 2.6265 218 0.7546 0.6159 0.7546 0.8687
No log 2.6506 220 0.7665 0.6070 0.7665 0.8755
No log 2.6747 222 0.7776 0.6376 0.7776 0.8818
No log 2.6988 224 0.8361 0.5456 0.8361 0.9144
No log 2.7229 226 0.8832 0.5235 0.8832 0.9398
No log 2.7470 228 0.8873 0.5245 0.8873 0.9420
No log 2.7711 230 0.7919 0.6269 0.7919 0.8899
No log 2.7952 232 0.7052 0.6617 0.7052 0.8398
No log 2.8193 234 0.6876 0.6527 0.6876 0.8292
No log 2.8434 236 0.7410 0.6509 0.7410 0.8608
No log 2.8675 238 0.7132 0.6565 0.7132 0.8445
No log 2.8916 240 0.6866 0.6600 0.6866 0.8286
No log 2.9157 242 0.6859 0.6521 0.6859 0.8282
No log 2.9398 244 0.7161 0.6598 0.7161 0.8462
No log 2.9639 246 0.7644 0.6486 0.7644 0.8743
No log 2.9880 248 0.8144 0.6572 0.8144 0.9024
No log 3.0120 250 0.7896 0.6549 0.7896 0.8886
No log 3.0361 252 0.7790 0.6578 0.7790 0.8826
No log 3.0602 254 0.7788 0.6402 0.7788 0.8825
No log 3.0843 256 0.7138 0.6730 0.7138 0.8448
No log 3.1084 258 0.7084 0.6556 0.7084 0.8417
No log 3.1325 260 0.8381 0.5968 0.8381 0.9155
No log 3.1566 262 1.0625 0.5250 1.0625 1.0308
No log 3.1807 264 1.0360 0.5111 1.0360 1.0178
No log 3.2048 266 0.8278 0.5748 0.8278 0.9098
No log 3.2289 268 0.7735 0.5681 0.7735 0.8795
No log 3.2530 270 0.8163 0.5311 0.8163 0.9035
No log 3.2771 272 0.8404 0.5079 0.8404 0.9167
No log 3.3012 274 0.8455 0.5242 0.8455 0.9195
No log 3.3253 276 0.8673 0.5802 0.8673 0.9313
No log 3.3494 278 0.9027 0.5808 0.9027 0.9501
No log 3.3735 280 0.8542 0.6183 0.8542 0.9242
No log 3.3976 282 0.7596 0.6525 0.7596 0.8716
No log 3.4217 284 0.7551 0.6885 0.7551 0.8690
No log 3.4458 286 0.7621 0.6728 0.7621 0.8730
No log 3.4699 288 0.7441 0.6721 0.7441 0.8626
No log 3.4940 290 0.7380 0.6803 0.7380 0.8591
No log 3.5181 292 0.7259 0.6547 0.7259 0.8520
No log 3.5422 294 0.7867 0.6196 0.7867 0.8870
No log 3.5663 296 0.8098 0.6470 0.8098 0.8999
No log 3.5904 298 0.8043 0.6607 0.8043 0.8968
No log 3.6145 300 0.7731 0.6509 0.7731 0.8793
No log 3.6386 302 0.8078 0.6482 0.8078 0.8988
No log 3.6627 304 0.8192 0.6549 0.8192 0.9051
No log 3.6867 306 0.8283 0.6471 0.8283 0.9101
No log 3.7108 308 0.8560 0.6551 0.8560 0.9252
No log 3.7349 310 0.8773 0.6365 0.8773 0.9366
No log 3.7590 312 0.8192 0.6514 0.8192 0.9051
No log 3.7831 314 0.8037 0.6953 0.8037 0.8965
No log 3.8072 316 0.7932 0.6882 0.7932 0.8906
No log 3.8313 318 0.8341 0.6727 0.8341 0.9133
No log 3.8554 320 0.8336 0.6655 0.8336 0.9130
No log 3.8795 322 0.8141 0.6252 0.8141 0.9023
No log 3.9036 324 0.8273 0.6054 0.8273 0.9096
No log 3.9277 326 0.9208 0.6050 0.9208 0.9596
No log 3.9518 328 1.0376 0.4692 1.0376 1.0186
No log 3.9759 330 1.2483 0.4340 1.2483 1.1173
No log 4.0 332 1.5774 0.3336 1.5774 1.2559
No log 4.0241 334 1.4517 0.3746 1.4517 1.2049
No log 4.0482 336 1.2806 0.4519 1.2806 1.1316
No log 4.0723 338 1.1181 0.4848 1.1181 1.0574
No log 4.0964 340 0.9994 0.5060 0.9994 0.9997
No log 4.1205 342 0.8830 0.5495 0.8830 0.9397
No log 4.1446 344 0.8276 0.5909 0.8276 0.9097
No log 4.1687 346 0.8032 0.6316 0.8032 0.8962
No log 4.1928 348 0.8321 0.6314 0.8321 0.9122
No log 4.2169 350 0.8892 0.6018 0.8892 0.9430
No log 4.2410 352 0.8281 0.6268 0.8281 0.9100
No log 4.2651 354 0.7441 0.6483 0.7441 0.8626
No log 4.2892 356 0.7031 0.6863 0.7031 0.8385
No log 4.3133 358 0.7270 0.6749 0.7270 0.8526
No log 4.3373 360 0.7202 0.6847 0.7202 0.8487
No log 4.3614 362 0.7193 0.6943 0.7193 0.8481
No log 4.3855 364 0.7088 0.6936 0.7088 0.8419
No log 4.4096 366 0.6958 0.6989 0.6958 0.8342
No log 4.4337 368 0.6641 0.7102 0.6641 0.8149
No log 4.4578 370 0.6485 0.6777 0.6485 0.8053
No log 4.4819 372 0.6764 0.6732 0.6764 0.8224
No log 4.5060 374 0.6786 0.6741 0.6786 0.8238
No log 4.5301 376 0.6714 0.6708 0.6714 0.8194
No log 4.5542 378 0.6794 0.6770 0.6794 0.8243
No log 4.5783 380 0.6792 0.6678 0.6792 0.8241
No log 4.6024 382 0.6772 0.6807 0.6772 0.8229
No log 4.6265 384 0.6822 0.6757 0.6822 0.8259
No log 4.6506 386 0.6829 0.6699 0.6829 0.8264
No log 4.6747 388 0.6760 0.6844 0.6760 0.8222
No log 4.6988 390 0.6992 0.6772 0.6992 0.8362
No log 4.7229 392 0.7258 0.6807 0.7258 0.8519
No log 4.7470 394 0.7138 0.6742 0.7138 0.8448
No log 4.7711 396 0.7199 0.6652 0.7199 0.8484
No log 4.7952 398 0.7079 0.6651 0.7079 0.8413
No log 4.8193 400 0.6984 0.6886 0.6984 0.8357
No log 4.8434 402 0.7086 0.6876 0.7086 0.8418
No log 4.8675 404 0.7148 0.6893 0.7148 0.8455
No log 4.8916 406 0.7179 0.6748 0.7179 0.8473
No log 4.9157 408 0.7204 0.6759 0.7204 0.8488
No log 4.9398 410 0.6787 0.6710 0.6787 0.8238
No log 4.9639 412 0.6592 0.6740 0.6592 0.8119
No log 4.9880 414 0.6632 0.6586 0.6632 0.8144
No log 5.0120 416 0.6692 0.6570 0.6692 0.8180
No log 5.0361 418 0.8066 0.6386 0.8066 0.8981
No log 5.0602 420 0.9186 0.6459 0.9186 0.9584
No log 5.0843 422 0.8347 0.6376 0.8347 0.9136
No log 5.1084 424 0.7254 0.6401 0.7254 0.8517
No log 5.1325 426 0.7009 0.6056 0.7009 0.8372
No log 5.1566 428 0.6990 0.6078 0.6990 0.8361
No log 5.1807 430 0.7408 0.6595 0.7408 0.8607
No log 5.2048 432 0.9012 0.6569 0.9012 0.9493
No log 5.2289 434 1.0429 0.5707 1.0429 1.0212
No log 5.2530 436 0.9414 0.6335 0.9414 0.9703
No log 5.2771 438 0.7411 0.6498 0.7411 0.8608
No log 5.3012 440 0.7064 0.6130 0.7064 0.8405
No log 5.3253 442 0.7064 0.6238 0.7064 0.8405
No log 5.3494 444 0.6810 0.6237 0.6810 0.8253
No log 5.3735 446 0.7634 0.6470 0.7634 0.8737
No log 5.3976 448 0.8395 0.6707 0.8395 0.9162
No log 5.4217 450 0.7828 0.6349 0.7828 0.8847
No log 5.4458 452 0.7232 0.6493 0.7232 0.8504
No log 5.4699 454 0.7130 0.6567 0.7130 0.8444
No log 5.4940 456 0.7364 0.6285 0.7364 0.8582
No log 5.5181 458 0.7403 0.6366 0.7403 0.8604
No log 5.5422 460 0.7759 0.6352 0.7759 0.8808
No log 5.5663 462 0.8899 0.5944 0.8899 0.9434
No log 5.5904 464 1.0416 0.5264 1.0416 1.0206
No log 5.6145 466 1.0190 0.4974 1.0190 1.0095
No log 5.6386 468 0.8751 0.5910 0.8751 0.9355
No log 5.6627 470 0.7570 0.6068 0.7570 0.8700
No log 5.6867 472 0.7530 0.6213 0.7530 0.8678
No log 5.7108 474 0.7747 0.6144 0.7747 0.8802
No log 5.7349 476 0.8036 0.6224 0.8036 0.8964
No log 5.7590 478 0.8679 0.5802 0.8679 0.9316
No log 5.7831 480 0.8234 0.6249 0.8234 0.9074
No log 5.8072 482 0.7494 0.6172 0.7494 0.8657
No log 5.8313 484 0.7127 0.6358 0.7127 0.8442
No log 5.8554 486 0.7079 0.6380 0.7079 0.8414
No log 5.8795 488 0.7485 0.6389 0.7485 0.8651
No log 5.9036 490 0.8226 0.6554 0.8226 0.9070
No log 5.9277 492 0.8631 0.6647 0.8631 0.9291
No log 5.9518 494 0.8884 0.6923 0.8884 0.9426
No log 5.9759 496 0.7741 0.6775 0.7741 0.8798
No log 6.0 498 0.6949 0.6869 0.6949 0.8336
0.4605 6.0241 500 0.6897 0.6866 0.6897 0.8305
0.4605 6.0482 502 0.7391 0.6497 0.7391 0.8597
0.4605 6.0723 504 0.8406 0.6677 0.8406 0.9168
0.4605 6.0964 506 0.8373 0.6716 0.8373 0.9150
0.4605 6.1205 508 0.7828 0.6507 0.7828 0.8848
0.4605 6.1446 510 0.7284 0.6739 0.7284 0.8534

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits8_usingALLEssays_FineTuningAraBERT_run1_AugV5_k17_task1_organization

Finetuned
(4023)
this model