ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k11_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7699
  • Qwk: 0.7273
  • Mse: 0.7699
  • Rmse: 0.8774

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0241 2 6.7222 0.0116 6.7222 2.5927
No log 0.0482 4 5.0076 0.0843 5.0076 2.2378
No log 0.0723 6 2.9427 0.0702 2.9427 1.7154
No log 0.0964 8 2.2130 0.1594 2.2130 1.4876
No log 0.1205 10 1.8152 0.1887 1.8152 1.3473
No log 0.1446 12 1.6864 0.1905 1.6864 1.2986
No log 0.1687 14 1.5282 0.1905 1.5282 1.2362
No log 0.1928 16 1.5493 0.2807 1.5493 1.2447
No log 0.2169 18 1.5136 0.3833 1.5136 1.2303
No log 0.2410 20 1.3462 0.4 1.3462 1.1603
No log 0.2651 22 1.0441 0.5 1.0441 1.0218
No log 0.2892 24 0.9816 0.544 0.9816 0.9908
No log 0.3133 26 0.9553 0.6667 0.9553 0.9774
No log 0.3373 28 1.2072 0.5211 1.2072 1.0987
No log 0.3614 30 1.1145 0.5793 1.1145 1.0557
No log 0.3855 32 1.2508 0.5541 1.2508 1.1184
No log 0.4096 34 1.1679 0.5850 1.1679 1.0807
No log 0.4337 36 0.6488 0.7248 0.6488 0.8055
No log 0.4578 38 0.8134 0.7081 0.8134 0.9019
No log 0.4819 40 0.7031 0.7619 0.7031 0.8385
No log 0.5060 42 0.7112 0.7730 0.7112 0.8433
No log 0.5301 44 0.8958 0.6871 0.8958 0.9465
No log 0.5542 46 0.7449 0.7643 0.7449 0.8631
No log 0.5783 48 0.5872 0.7771 0.5872 0.7663
No log 0.6024 50 0.8630 0.6795 0.8630 0.9290
No log 0.6265 52 0.9374 0.6447 0.9374 0.9682
No log 0.6506 54 0.7443 0.6933 0.7443 0.8627
No log 0.6747 56 0.7243 0.7517 0.7243 0.8511
No log 0.6988 58 0.7723 0.7042 0.7723 0.8788
No log 0.7229 60 0.7523 0.7465 0.7523 0.8673
No log 0.7470 62 0.7583 0.7324 0.7583 0.8708
No log 0.7711 64 0.7550 0.6993 0.7550 0.8689
No log 0.7952 66 0.7497 0.7671 0.7497 0.8658
No log 0.8193 68 0.9496 0.6241 0.9496 0.9745
No log 0.8434 70 0.9632 0.5735 0.9632 0.9814
No log 0.8675 72 0.7503 0.6897 0.7503 0.8662
No log 0.8916 74 0.8516 0.6974 0.8516 0.9228
No log 0.9157 76 0.7741 0.7027 0.7741 0.8798
No log 0.9398 78 0.8130 0.7172 0.8130 0.9017
No log 0.9639 80 0.7993 0.7133 0.7993 0.8940
No log 0.9880 82 0.7436 0.7211 0.7436 0.8623
No log 1.0120 84 0.7618 0.7273 0.7618 0.8728
No log 1.0361 86 0.7430 0.7067 0.7430 0.8620
No log 1.0602 88 0.7615 0.7361 0.7615 0.8726
No log 1.0843 90 0.7741 0.7067 0.7741 0.8799
No log 1.1084 92 0.8105 0.7190 0.8105 0.9003
No log 1.1325 94 0.7628 0.7248 0.7628 0.8734
No log 1.1566 96 0.7975 0.7260 0.7975 0.8930
No log 1.1807 98 0.7950 0.7342 0.7950 0.8916
No log 1.2048 100 0.7358 0.7439 0.7358 0.8578
No log 1.2289 102 0.6977 0.7531 0.6977 0.8353
No log 1.2530 104 0.7267 0.7442 0.7267 0.8525
No log 1.2771 106 0.8360 0.7447 0.8360 0.9143
No log 1.3012 108 0.9722 0.7451 0.9722 0.9860
No log 1.3253 110 0.8960 0.76 0.8960 0.9466
No log 1.3494 112 0.8876 0.7351 0.8876 0.9421
No log 1.3735 114 0.8210 0.7101 0.8210 0.9061
No log 1.3976 116 0.7448 0.7226 0.7448 0.8630
No log 1.4217 118 0.7522 0.7467 0.7522 0.8673
No log 1.4458 120 0.7152 0.7619 0.7152 0.8457
No log 1.4699 122 0.7107 0.7027 0.7107 0.8430
No log 1.4940 124 0.7284 0.7260 0.7284 0.8535
No log 1.5181 126 0.7373 0.7172 0.7373 0.8587
No log 1.5422 128 0.7707 0.6986 0.7707 0.8779
No log 1.5663 130 0.7712 0.7297 0.7712 0.8782
No log 1.5904 132 0.7540 0.7123 0.7540 0.8683
No log 1.6145 134 0.7365 0.7285 0.7365 0.8582
No log 1.6386 136 0.7847 0.7083 0.7847 0.8859
No log 1.6627 138 0.7629 0.7211 0.7629 0.8734
No log 1.6867 140 0.7237 0.7517 0.7237 0.8507
No log 1.7108 142 0.7548 0.7451 0.7548 0.8688
No log 1.7349 144 0.7273 0.7368 0.7273 0.8528
No log 1.7590 146 0.9369 0.7093 0.9369 0.9679
No log 1.7831 148 1.0898 0.6845 1.0898 1.0439
No log 1.8072 150 0.8959 0.7030 0.8959 0.9465
No log 1.8313 152 0.7686 0.7179 0.7686 0.8767
No log 1.8554 154 0.8680 0.6871 0.8680 0.9317
No log 1.8795 156 1.1248 0.6258 1.1248 1.0605
No log 1.9036 158 1.0072 0.6585 1.0072 1.0036
No log 1.9277 160 0.7460 0.7389 0.7460 0.8637
No log 1.9518 162 0.7570 0.7317 0.7570 0.8700
No log 1.9759 164 0.7983 0.7262 0.7983 0.8935
No log 2.0 166 0.7207 0.7190 0.7207 0.8489
No log 2.0241 168 0.7277 0.7467 0.7277 0.8531
No log 2.0482 170 0.7488 0.7383 0.7488 0.8653
No log 2.0723 172 0.7280 0.7397 0.7280 0.8532
No log 2.0964 174 0.6878 0.7413 0.6878 0.8294
No log 2.1205 176 0.6685 0.7724 0.6685 0.8176
No log 2.1446 178 0.7018 0.7712 0.7018 0.8378
No log 2.1687 180 0.8272 0.6755 0.8272 0.9095
No log 2.1928 182 0.8661 0.6579 0.8661 0.9307
No log 2.2169 184 0.7530 0.7237 0.7530 0.8678
No log 2.2410 186 0.7120 0.8050 0.7120 0.8438
No log 2.2651 188 0.8277 0.6790 0.8277 0.9098
No log 2.2892 190 0.7166 0.7439 0.7166 0.8465
No log 2.3133 192 0.6551 0.8313 0.6551 0.8094
No log 2.3373 194 0.6928 0.7848 0.6928 0.8324
No log 2.3614 196 0.7059 0.7815 0.7059 0.8401
No log 2.3855 198 0.7183 0.7733 0.7183 0.8475
No log 2.4096 200 0.6984 0.7092 0.6984 0.8357
No log 2.4337 202 0.6821 0.7133 0.6821 0.8259
No log 2.4578 204 0.6675 0.7260 0.6675 0.8170
No log 2.4819 206 0.6372 0.7826 0.6372 0.7983
No log 2.5060 208 0.6855 0.7886 0.6855 0.8280
No log 2.5301 210 0.7265 0.7442 0.7265 0.8524
No log 2.5542 212 0.7410 0.75 0.7410 0.8608
No log 2.5783 214 0.7868 0.7397 0.7868 0.8870
No log 2.6024 216 0.7384 0.7075 0.7384 0.8593
No log 2.6265 218 0.6880 0.7485 0.6880 0.8294
No log 2.6506 220 0.7367 0.7514 0.7367 0.8583
No log 2.6747 222 0.6728 0.7674 0.6728 0.8203
No log 2.6988 224 0.6299 0.7952 0.6299 0.7936
No log 2.7229 226 0.6048 0.7578 0.6048 0.7777
No log 2.7470 228 0.6637 0.7952 0.6637 0.8146
No log 2.7711 230 0.7274 0.72 0.7274 0.8529
No log 2.7952 232 0.7188 0.75 0.7188 0.8478
No log 2.8193 234 0.7812 0.6619 0.7812 0.8839
No log 2.8434 236 0.8673 0.6667 0.8673 0.9313
No log 2.8675 238 0.8013 0.7261 0.8013 0.8951
No log 2.8916 240 0.6694 0.7421 0.6694 0.8181
No log 2.9157 242 0.6627 0.7799 0.6627 0.8141
No log 2.9398 244 0.7919 0.7665 0.7919 0.8899
No log 2.9639 246 0.7323 0.7882 0.7323 0.8558
No log 2.9880 248 0.6369 0.7647 0.6369 0.7981
No log 3.0120 250 0.9341 0.7273 0.9341 0.9665
No log 3.0361 252 1.1834 0.6344 1.1834 1.0878
No log 3.0602 254 1.0755 0.6386 1.0755 1.0371
No log 3.0843 256 0.8985 0.6763 0.8985 0.9479
No log 3.1084 258 0.8274 0.6716 0.8274 0.9096
No log 3.1325 260 0.7793 0.6763 0.7793 0.8828
No log 3.1566 262 0.7126 0.7027 0.7126 0.8442
No log 3.1807 264 0.6620 0.7682 0.6620 0.8137
No log 3.2048 266 0.6667 0.7407 0.6667 0.8165
No log 3.2289 268 0.7975 0.7241 0.7975 0.8930
No log 3.2530 270 0.8734 0.7263 0.8734 0.9345
No log 3.2771 272 0.8528 0.7059 0.8528 0.9235
No log 3.3012 274 0.7203 0.7261 0.7203 0.8487
No log 3.3253 276 0.6796 0.7755 0.6796 0.8244
No log 3.3494 278 0.6888 0.7534 0.6888 0.8299
No log 3.3735 280 0.7438 0.7222 0.7438 0.8624
No log 3.3976 282 0.7876 0.7333 0.7876 0.8875
No log 3.4217 284 0.7288 0.7248 0.7288 0.8537
No log 3.4458 286 0.6445 0.7606 0.6445 0.8028
No log 3.4699 288 0.6418 0.7286 0.6418 0.8011
No log 3.4940 290 0.6224 0.75 0.6224 0.7889
No log 3.5181 292 0.6332 0.7826 0.6332 0.7958
No log 3.5422 294 0.6475 0.7682 0.6475 0.8047
No log 3.5663 296 0.6941 0.7445 0.6941 0.8332
No log 3.5904 298 0.7310 0.7246 0.7310 0.8550
No log 3.6145 300 0.7624 0.7013 0.7624 0.8731
No log 3.6386 302 0.6998 0.7375 0.6998 0.8366
No log 3.6627 304 0.6071 0.7260 0.6071 0.7791
No log 3.6867 306 0.5765 0.7417 0.5765 0.7593
No log 3.7108 308 0.5937 0.775 0.5937 0.7705
No log 3.7349 310 0.6583 0.7636 0.6583 0.8113
No log 3.7590 312 0.6704 0.7665 0.6704 0.8188
No log 3.7831 314 0.6138 0.7590 0.6138 0.7834
No log 3.8072 316 0.5994 0.7517 0.5994 0.7742
No log 3.8313 318 0.5540 0.8054 0.5540 0.7443
No log 3.8554 320 0.5293 0.7919 0.5293 0.7275
No log 3.8795 322 0.5120 0.8302 0.5120 0.7155
No log 3.9036 324 0.5413 0.8079 0.5413 0.7358
No log 3.9277 326 0.5569 0.7891 0.5569 0.7463
No log 3.9518 328 0.5727 0.8 0.5727 0.7568
No log 3.9759 330 0.6136 0.7901 0.6136 0.7833
No log 4.0 332 0.6362 0.75 0.6362 0.7976
No log 4.0241 334 0.6825 0.7568 0.6825 0.8262
No log 4.0482 336 0.6936 0.7183 0.6936 0.8328
No log 4.0723 338 0.7242 0.7183 0.7242 0.8510
No log 4.0964 340 0.8594 0.6713 0.8594 0.9270
No log 4.1205 342 0.8774 0.6569 0.8774 0.9367
No log 4.1446 344 0.8051 0.6519 0.8051 0.8973
No log 4.1687 346 0.7282 0.7164 0.7282 0.8533
No log 4.1928 348 0.8276 0.6879 0.8276 0.9097
No log 4.2169 350 0.8604 0.6826 0.8604 0.9276
No log 4.2410 352 0.7414 0.7179 0.7414 0.8611
No log 4.2651 354 0.6267 0.7857 0.6267 0.7917
No log 4.2892 356 0.6659 0.75 0.6659 0.8160
No log 4.3133 358 0.6596 0.75 0.6596 0.8122
No log 4.3373 360 0.6304 0.7917 0.6304 0.7940
No log 4.3614 362 0.6748 0.7925 0.6748 0.8215
No log 4.3855 364 0.8169 0.7051 0.8169 0.9038
No log 4.4096 366 0.8413 0.6712 0.8413 0.9172
No log 4.4337 368 0.8029 0.6912 0.8029 0.8960
No log 4.4578 370 0.7921 0.6912 0.7921 0.8900
No log 4.4819 372 0.7365 0.7285 0.7365 0.8582
No log 4.5060 374 0.6381 0.7662 0.6381 0.7988
No log 4.5301 376 0.5976 0.8144 0.5976 0.7731
No log 4.5542 378 0.6333 0.8 0.6333 0.7958
No log 4.5783 380 0.6482 0.7811 0.6482 0.8051
No log 4.6024 382 0.6498 0.7532 0.6498 0.8061
No log 4.6265 384 0.6738 0.7413 0.6738 0.8209
No log 4.6506 386 0.6300 0.7550 0.6300 0.7937
No log 4.6747 388 0.5472 0.8171 0.5472 0.7397
No log 4.6988 390 0.5044 0.8457 0.5044 0.7102
No log 4.7229 392 0.5672 0.8105 0.5672 0.7531
No log 4.7470 394 0.6543 0.7872 0.6543 0.8089
No log 4.7711 396 0.5795 0.8065 0.5795 0.7612
No log 4.7952 398 0.4813 0.8136 0.4813 0.6938
No log 4.8193 400 0.5246 0.8295 0.5246 0.7243
No log 4.8434 402 0.5953 0.7673 0.5953 0.7716
No log 4.8675 404 0.6593 0.7143 0.6593 0.8120
No log 4.8916 406 0.6770 0.7183 0.6770 0.8228
No log 4.9157 408 0.6530 0.7518 0.6530 0.8081
No log 4.9398 410 0.6749 0.7183 0.6749 0.8215
No log 4.9639 412 0.7046 0.7647 0.7046 0.8394
No log 4.9880 414 0.8031 0.7429 0.8031 0.8962
No log 5.0120 416 0.8310 0.7273 0.8310 0.9116
No log 5.0361 418 0.7198 0.7368 0.7198 0.8484
No log 5.0602 420 0.6008 0.7947 0.6008 0.7751
No log 5.0843 422 0.6224 0.7606 0.6224 0.7889
No log 5.1084 424 0.6285 0.7552 0.6285 0.7928
No log 5.1325 426 0.6158 0.7626 0.6158 0.7847
No log 5.1566 428 0.6954 0.7020 0.6954 0.8339
No log 5.1807 430 0.8606 0.7219 0.8606 0.9277
No log 5.2048 432 0.9841 0.6429 0.9841 0.9920
No log 5.2289 434 0.9489 0.6946 0.9489 0.9741
No log 5.2530 436 0.8115 0.6667 0.8115 0.9008
No log 5.2771 438 0.6915 0.7778 0.6915 0.8315
No log 5.3012 440 0.6038 0.8194 0.6038 0.7770
No log 5.3253 442 0.5586 0.8188 0.5586 0.7474
No log 5.3494 444 0.5289 0.8176 0.5289 0.7273
No log 5.3735 446 0.5285 0.8176 0.5285 0.7270
No log 5.3976 448 0.5618 0.8105 0.5618 0.7495
No log 5.4217 450 0.6040 0.7947 0.6040 0.7772
No log 5.4458 452 0.6333 0.7376 0.6333 0.7958
No log 5.4699 454 0.6829 0.7153 0.6829 0.8264
No log 5.4940 456 0.7406 0.7259 0.7406 0.8606
No log 5.5181 458 0.8104 0.7121 0.8104 0.9002
No log 5.5422 460 0.8228 0.6923 0.8228 0.9071
No log 5.5663 462 0.7845 0.7121 0.7845 0.8857
No log 5.5904 464 0.7207 0.7111 0.7207 0.8489
No log 5.6145 466 0.6684 0.7273 0.6684 0.8176
No log 5.6386 468 0.6261 0.7895 0.6261 0.7913
No log 5.6627 470 0.5955 0.8108 0.5955 0.7717
No log 5.6867 472 0.5869 0.8025 0.5869 0.7661
No log 5.7108 474 0.5809 0.8176 0.5809 0.7622
No log 5.7349 476 0.5863 0.7871 0.5863 0.7657
No log 5.7590 478 0.6250 0.7843 0.6250 0.7906
No log 5.7831 480 0.7432 0.6957 0.7432 0.8621
No log 5.8072 482 0.8689 0.6626 0.8689 0.9321
No log 5.8313 484 0.8375 0.6494 0.8375 0.9152
No log 5.8554 486 0.7333 0.7183 0.7333 0.8563
No log 5.8795 488 0.6673 0.7746 0.6673 0.8169
No log 5.9036 490 0.6590 0.7724 0.6590 0.8118
No log 5.9277 492 0.6333 0.8 0.6333 0.7958
No log 5.9518 494 0.6175 0.7867 0.6175 0.7858
No log 5.9759 496 0.7322 0.7283 0.7322 0.8557
No log 6.0 498 0.8832 0.7232 0.8832 0.9398
0.3923 6.0241 500 0.9356 0.7273 0.9356 0.9672
0.3923 6.0482 502 0.8405 0.7143 0.8405 0.9168
0.3923 6.0723 504 0.7195 0.6968 0.7195 0.8482
0.3923 6.0964 506 0.6735 0.7919 0.6735 0.8207
0.3923 6.1205 508 0.6941 0.7778 0.6941 0.8331
0.3923 6.1446 510 0.7074 0.7778 0.7074 0.8411
0.3923 6.1687 512 0.7012 0.7619 0.7012 0.8374
0.3923 6.1928 514 0.7560 0.7092 0.7560 0.8695
0.3923 6.2169 516 0.8323 0.6667 0.8323 0.9123
0.3923 6.2410 518 0.8287 0.6853 0.8287 0.9103
0.3923 6.2651 520 0.7898 0.6812 0.7898 0.8887
0.3923 6.2892 522 0.7699 0.7273 0.7699 0.8774

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k11_task1_organization

Finetuned
(4023)
this model