ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k8_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6515
  • Qwk: 0.7949
  • Mse: 0.6515
  • Rmse: 0.8071

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 7.2060 0.0 7.2060 2.6844
No log 0.1053 4 4.7106 0.0606 4.7106 2.1704
No log 0.1579 6 3.1661 0.0833 3.1661 1.7794
No log 0.2105 8 2.9879 0.0513 2.9879 1.7285
No log 0.2632 10 2.1518 0.2059 2.1518 1.4669
No log 0.3158 12 1.8024 0.3443 1.8024 1.3425
No log 0.3684 14 1.9070 0.3101 1.9070 1.3809
No log 0.4211 16 1.6108 0.3390 1.6108 1.2692
No log 0.4737 18 1.7225 0.3492 1.7225 1.3124
No log 0.5263 20 1.6669 0.3594 1.6669 1.2911
No log 0.5789 22 1.6720 0.3731 1.6720 1.2931
No log 0.6316 24 1.8449 0.3889 1.8449 1.3583
No log 0.6842 26 1.6312 0.4265 1.6312 1.2772
No log 0.7368 28 1.3728 0.4844 1.3728 1.1717
No log 0.7895 30 1.3892 0.4923 1.3892 1.1787
No log 0.8421 32 2.1746 0.2857 2.1746 1.4747
No log 0.8947 34 2.7118 0.2024 2.7118 1.6467
No log 0.9474 36 2.3050 0.2222 2.3050 1.5182
No log 1.0 38 1.4400 0.4769 1.4400 1.2000
No log 1.0526 40 1.1237 0.5669 1.1237 1.0600
No log 1.1053 42 1.2077 0.4882 1.2077 1.0990
No log 1.1579 44 1.5715 0.4031 1.5715 1.2536
No log 1.2105 46 2.0606 0.1418 2.0606 1.4355
No log 1.2632 48 2.6033 0.1290 2.6033 1.6135
No log 1.3158 50 2.6088 0.1290 2.6088 1.6152
No log 1.3684 52 2.2544 0.1830 2.2544 1.5015
No log 1.4211 54 1.5883 0.4681 1.5883 1.2603
No log 1.4737 56 1.1228 0.5263 1.1228 1.0596
No log 1.5263 58 1.0903 0.5496 1.0903 1.0442
No log 1.5789 60 1.0978 0.5672 1.0978 1.0478
No log 1.6316 62 1.0118 0.6479 1.0118 1.0059
No log 1.6842 64 0.9486 0.6712 0.9486 0.9740
No log 1.7368 66 0.9257 0.7114 0.9257 0.9621
No log 1.7895 68 0.7919 0.7285 0.7919 0.8899
No log 1.8421 70 0.8092 0.7114 0.8092 0.8996
No log 1.8947 72 0.8540 0.7027 0.8540 0.9241
No log 1.9474 74 0.7533 0.7114 0.7533 0.8679
No log 2.0 76 0.7724 0.7682 0.7724 0.8788
No log 2.0526 78 1.2669 0.5930 1.2669 1.1256
No log 2.1053 80 1.2972 0.5943 1.2972 1.1389
No log 2.1579 82 1.0241 0.6588 1.0241 1.0120
No log 2.2105 84 0.7653 0.7826 0.7653 0.8748
No log 2.2632 86 0.8088 0.6806 0.8088 0.8993
No log 2.3158 88 0.8542 0.6573 0.8542 0.9242
No log 2.3684 90 0.7670 0.7733 0.7670 0.8758
No log 2.4211 92 0.8885 0.6857 0.8885 0.9426
No log 2.4737 94 0.9279 0.6588 0.9279 0.9633
No log 2.5263 96 0.7735 0.7711 0.7735 0.8795
No log 2.5789 98 0.7497 0.72 0.7497 0.8659
No log 2.6316 100 0.9911 0.6806 0.9911 0.9955
No log 2.6842 102 0.9364 0.6806 0.9364 0.9677
No log 2.7368 104 0.8455 0.7027 0.8455 0.9195
No log 2.7895 106 0.8393 0.7368 0.8393 0.9161
No log 2.8421 108 0.8755 0.6575 0.8755 0.9357
No log 2.8947 110 0.8117 0.7105 0.8117 0.9009
No log 2.9474 112 0.8686 0.7125 0.8686 0.9320
No log 3.0 114 0.9616 0.7006 0.9616 0.9806
No log 3.0526 116 0.9520 0.7226 0.9520 0.9757
No log 3.1053 118 0.9811 0.6056 0.9811 0.9905
No log 3.1579 120 0.9783 0.6154 0.9783 0.9891
No log 3.2105 122 0.8760 0.6806 0.8760 0.9359
No log 3.2632 124 0.9050 0.7011 0.9050 0.9513
No log 3.3158 126 1.1717 0.5914 1.1717 1.0824
No log 3.3684 128 1.1984 0.5882 1.1984 1.0947
No log 3.4211 130 0.8507 0.7086 0.8507 0.9223
No log 3.4737 132 0.7916 0.7484 0.7916 0.8897
No log 3.5263 134 1.0113 0.6351 1.0113 1.0056
No log 3.5789 136 1.0035 0.6531 1.0035 1.0017
No log 3.6316 138 0.9671 0.6533 0.9671 0.9834
No log 3.6842 140 1.0797 0.6108 1.0797 1.0391
No log 3.7368 142 1.1291 0.6061 1.1291 1.0626
No log 3.7895 144 0.9738 0.6842 0.9738 0.9868
No log 3.8421 146 0.8026 0.7792 0.8026 0.8959
No log 3.8947 148 0.7708 0.7815 0.7708 0.8779
No log 3.9474 150 0.7165 0.7733 0.7165 0.8465
No log 4.0 152 0.6810 0.7733 0.6810 0.8252
No log 4.0526 154 0.6784 0.7517 0.6784 0.8236
No log 4.1053 156 0.6654 0.7467 0.6654 0.8157
No log 4.1579 158 0.7286 0.7248 0.7286 0.8536
No log 4.2105 160 0.7936 0.6849 0.7936 0.8908
No log 4.2632 162 0.7921 0.7075 0.7921 0.8900
No log 4.3158 164 0.7772 0.7550 0.7772 0.8816
No log 4.3684 166 0.8047 0.7662 0.8047 0.8970
No log 4.4211 168 0.7755 0.7662 0.7755 0.8806
No log 4.4737 170 0.7697 0.7568 0.7697 0.8773
No log 4.5263 172 0.7556 0.7075 0.7556 0.8693
No log 4.5789 174 0.7006 0.7297 0.7006 0.8370
No log 4.6316 176 0.6462 0.7815 0.6462 0.8039
No log 4.6842 178 0.6520 0.8025 0.6520 0.8075
No log 4.7368 180 0.6350 0.8025 0.6350 0.7969
No log 4.7895 182 0.6329 0.8098 0.6329 0.7955
No log 4.8421 184 0.6295 0.8025 0.6295 0.7934
No log 4.8947 186 0.6472 0.7975 0.6472 0.8045
No log 4.9474 188 0.6756 0.7843 0.6756 0.8220
No log 5.0 190 0.7107 0.7662 0.7107 0.8430
No log 5.0526 192 0.6855 0.7950 0.6855 0.8280
No log 5.1053 194 0.7144 0.7975 0.7144 0.8452
No log 5.1579 196 0.7603 0.7602 0.7603 0.8720
No log 5.2105 198 0.7700 0.7485 0.7700 0.8775
No log 5.2632 200 0.6784 0.7975 0.6784 0.8237
No log 5.3158 202 0.6668 0.7619 0.6668 0.8166
No log 5.3684 204 0.6728 0.7619 0.6728 0.8203
No log 5.4211 206 0.6655 0.7791 0.6655 0.8158
No log 5.4737 208 0.8036 0.7232 0.8036 0.8965
No log 5.5263 210 0.7868 0.7556 0.7868 0.8870
No log 5.5789 212 0.6953 0.7799 0.6953 0.8339
No log 5.6316 214 0.8099 0.6846 0.8099 0.9000
No log 5.6842 216 1.0747 0.6014 1.0747 1.0367
No log 5.7368 218 1.1868 0.5755 1.1868 1.0894
No log 5.7895 220 1.1173 0.6277 1.1173 1.0570
No log 5.8421 222 1.1852 0.4531 1.1852 1.0887
No log 5.8947 224 1.1010 0.5156 1.1010 1.0493
No log 5.9474 226 1.0386 0.5156 1.0386 1.0191
No log 6.0 228 0.9410 0.6567 0.9410 0.9701
No log 6.0526 230 0.9344 0.6617 0.9344 0.9666
No log 6.1053 232 0.9289 0.6423 0.9289 0.9638
No log 6.1579 234 0.7965 0.7042 0.7965 0.8925
No log 6.2105 236 0.6901 0.7448 0.6901 0.8307
No log 6.2632 238 0.6901 0.7361 0.6901 0.8307
No log 6.3158 240 0.8191 0.6849 0.8191 0.9050
No log 6.3684 242 0.8838 0.6056 0.8838 0.9401
No log 6.4211 244 0.8688 0.6099 0.8688 0.9321
No log 6.4737 246 0.7308 0.7273 0.7308 0.8549
No log 6.5263 248 0.6699 0.7755 0.6699 0.8185
No log 6.5789 250 0.6785 0.7848 0.6785 0.8237
No log 6.6316 252 0.7239 0.7927 0.7239 0.8508
No log 6.6842 254 0.8286 0.7368 0.8286 0.9102
No log 6.7368 256 0.8822 0.7176 0.8822 0.9392
No log 6.7895 258 0.8175 0.7059 0.8175 0.9041
No log 6.8421 260 0.7894 0.7133 0.7894 0.8885
No log 6.8947 262 0.7630 0.7448 0.7630 0.8735
No log 6.9474 264 0.7058 0.7619 0.7058 0.8401
No log 7.0 266 0.6240 0.7733 0.6240 0.7899
No log 7.0526 268 0.5895 0.7815 0.5895 0.7678
No log 7.1053 270 0.6184 0.7722 0.6184 0.7864
No log 7.1579 272 0.5998 0.7682 0.5998 0.7744
No log 7.2105 274 0.6332 0.7651 0.6332 0.7957
No log 7.2632 276 0.6913 0.7376 0.6913 0.8314
No log 7.3158 278 0.6692 0.7552 0.6692 0.8180
No log 7.3684 280 0.6939 0.7534 0.6939 0.8330
No log 7.4211 282 0.8445 0.6887 0.8445 0.9189
No log 7.4737 284 0.9570 0.6405 0.9570 0.9783
No log 7.5263 286 0.8817 0.7051 0.8817 0.9390
No log 7.5789 288 0.7625 0.7320 0.7625 0.8732
No log 7.6316 290 0.6664 0.7619 0.6664 0.8163
No log 7.6842 292 0.6819 0.7534 0.6819 0.8258
No log 7.7368 294 0.7017 0.7448 0.7017 0.8377
No log 7.7895 296 0.7015 0.7586 0.7015 0.8376
No log 7.8421 298 0.7288 0.7436 0.7288 0.8537
No log 7.8947 300 0.7226 0.7643 0.7226 0.8500
No log 7.9474 302 0.6966 0.7975 0.6966 0.8346
No log 8.0 304 0.6799 0.7763 0.6799 0.8245
No log 8.0526 306 0.7379 0.7397 0.7379 0.8590
No log 8.1053 308 0.7998 0.7448 0.7998 0.8943
No log 8.1579 310 0.7727 0.7397 0.7727 0.8790
No log 8.2105 312 0.7263 0.7742 0.7263 0.8523
No log 8.2632 314 0.7877 0.7950 0.7877 0.8875
No log 8.3158 316 0.8386 0.7712 0.8386 0.9157
No log 8.3684 318 0.8498 0.7361 0.8498 0.9219
No log 8.4211 320 0.8147 0.7483 0.8147 0.9026
No log 8.4737 322 0.7825 0.7534 0.7825 0.8846
No log 8.5263 324 0.7463 0.7568 0.7463 0.8639
No log 8.5789 326 0.7088 0.7662 0.7088 0.8419
No log 8.6316 328 0.7125 0.7904 0.7125 0.8441
No log 8.6842 330 0.6542 0.8072 0.6542 0.8088
No log 8.7368 332 0.6509 0.7925 0.6509 0.8068
No log 8.7895 334 0.7075 0.7651 0.7075 0.8411
No log 8.8421 336 0.7763 0.7568 0.7763 0.8811
No log 8.8947 338 0.8250 0.7682 0.8250 0.9083
No log 8.9474 340 0.9107 0.6752 0.9107 0.9543
No log 9.0 342 0.9335 0.6296 0.9335 0.9662
No log 9.0526 344 0.8616 0.7673 0.8616 0.9282
No log 9.1053 346 0.7903 0.7568 0.7903 0.8890
No log 9.1579 348 0.7356 0.7651 0.7356 0.8577
No log 9.2105 350 0.6984 0.7632 0.6984 0.8357
No log 9.2632 352 0.6969 0.7901 0.6969 0.8348
No log 9.3158 354 0.6815 0.7568 0.6815 0.8255
No log 9.3684 356 0.6798 0.7432 0.6798 0.8245
No log 9.4211 358 0.7132 0.7397 0.7132 0.8445
No log 9.4737 360 0.7461 0.7172 0.7461 0.8638
No log 9.5263 362 0.7757 0.7172 0.7757 0.8807
No log 9.5789 364 0.7661 0.7172 0.7661 0.8753
No log 9.6316 366 0.7358 0.7123 0.7358 0.8578
No log 9.6842 368 0.7402 0.7123 0.7402 0.8603
No log 9.7368 370 0.7438 0.7310 0.7438 0.8625
No log 9.7895 372 0.7800 0.7448 0.7800 0.8832
No log 9.8421 374 0.8180 0.7075 0.8180 0.9044
No log 9.8947 376 0.7879 0.7034 0.7879 0.8876
No log 9.9474 378 0.7338 0.7534 0.7338 0.8566
No log 10.0 380 0.6678 0.7703 0.6678 0.8172
No log 10.0526 382 0.6630 0.7922 0.6630 0.8142
No log 10.1053 384 0.7426 0.7771 0.7426 0.8617
No log 10.1579 386 0.8408 0.7821 0.8408 0.9170
No log 10.2105 388 0.8384 0.7889 0.8384 0.9156
No log 10.2632 390 0.7508 0.7821 0.7508 0.8665
No log 10.3158 392 0.7667 0.7701 0.7667 0.8756
No log 10.3684 394 0.7757 0.7683 0.7757 0.8807
No log 10.4211 396 0.8319 0.7285 0.8319 0.9121
No log 10.4737 398 0.8444 0.7248 0.8444 0.9189
No log 10.5263 400 0.8695 0.7114 0.8695 0.9325
No log 10.5789 402 0.8746 0.7114 0.8746 0.9352
No log 10.6316 404 0.8732 0.7308 0.8732 0.9344
No log 10.6842 406 0.8591 0.7362 0.8591 0.9269
No log 10.7368 408 0.8537 0.7308 0.8537 0.9239
No log 10.7895 410 0.8544 0.7226 0.8544 0.9243
No log 10.8421 412 0.8447 0.7034 0.8447 0.9191
No log 10.8947 414 0.8240 0.7260 0.8240 0.9078
No log 10.9474 416 0.7966 0.7260 0.7966 0.8925
No log 11.0 418 0.7477 0.7273 0.7477 0.8647
No log 11.0526 420 0.7250 0.7083 0.7250 0.8514
No log 11.1053 422 0.7084 0.7222 0.7084 0.8416
No log 11.1579 424 0.7232 0.7260 0.7232 0.8504
No log 11.2105 426 0.7328 0.7260 0.7328 0.8560
No log 11.2632 428 0.7160 0.7172 0.7160 0.8462
No log 11.3158 430 0.7272 0.7413 0.7272 0.8528
No log 11.3684 432 0.7947 0.7260 0.7947 0.8914
No log 11.4211 434 0.8237 0.7248 0.8237 0.9076
No log 11.4737 436 0.7600 0.7467 0.7600 0.8718
No log 11.5263 438 0.7341 0.7582 0.7341 0.8568
No log 11.5789 440 0.7076 0.8077 0.7076 0.8412
No log 11.6316 442 0.6933 0.8153 0.6933 0.8326
No log 11.6842 444 0.7265 0.7625 0.7265 0.8524
No log 11.7368 446 0.7584 0.75 0.7584 0.8709
No log 11.7895 448 0.7548 0.75 0.7548 0.8688
No log 11.8421 450 0.6984 0.7871 0.6984 0.8357
No log 11.8947 452 0.6615 0.7974 0.6615 0.8133
No log 11.9474 454 0.6472 0.7843 0.6472 0.8045
No log 12.0 456 0.6495 0.8050 0.6495 0.8059
No log 12.0526 458 0.6778 0.8144 0.6778 0.8233
No log 12.1053 460 0.6906 0.8144 0.6906 0.8310
No log 12.1579 462 0.6907 0.7821 0.6907 0.8311
No log 12.2105 464 0.7289 0.7922 0.7289 0.8537
No log 12.2632 466 0.7237 0.7922 0.7237 0.8507
No log 12.3158 468 0.7075 0.7843 0.7075 0.8411
No log 12.3684 470 0.6963 0.7922 0.6963 0.8344
No log 12.4211 472 0.6782 0.8049 0.6782 0.8235
No log 12.4737 474 0.6688 0.7816 0.6688 0.8178
No log 12.5263 476 0.6786 0.7816 0.6786 0.8238
No log 12.5789 478 0.6764 0.7816 0.6764 0.8224
No log 12.6316 480 0.7203 0.7746 0.7203 0.8487
No log 12.6842 482 0.7436 0.7425 0.7436 0.8623
No log 12.7368 484 0.8138 0.7578 0.8138 0.9021
No log 12.7895 486 0.8530 0.6939 0.8530 0.9236
No log 12.8421 488 0.8941 0.6331 0.8941 0.9456
No log 12.8947 490 0.9021 0.6176 0.9021 0.9498
No log 12.9474 492 0.8346 0.7133 0.8346 0.9135
No log 13.0 494 0.7418 0.7568 0.7418 0.8613
No log 13.0526 496 0.6724 0.8052 0.6724 0.8200
No log 13.1053 498 0.6432 0.7975 0.6432 0.8020
0.3722 13.1579 500 0.6420 0.7976 0.6420 0.8013
0.3722 13.2105 502 0.6544 0.7976 0.6544 0.8089
0.3722 13.2632 504 0.6458 0.7952 0.6458 0.8036
0.3722 13.3158 506 0.6392 0.7949 0.6392 0.7995
0.3722 13.3684 508 0.6370 0.7843 0.6370 0.7981
0.3722 13.4211 510 0.6515 0.7949 0.6515 0.8071

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k8_task1_organization

Finetuned
(4023)
this model