ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k7_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9557
  • Qwk: 0.5891
  • Mse: 0.9557
  • Rmse: 0.9776

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0625 2 6.8134 0.0242 6.8134 2.6103
No log 0.125 4 4.5193 0.0803 4.5193 2.1259
No log 0.1875 6 3.1426 0.0714 3.1426 1.7727
No log 0.25 8 3.1190 0.0256 3.1190 1.7661
No log 0.3125 10 2.2591 0.1069 2.2591 1.5030
No log 0.375 12 1.9049 0.2167 1.9049 1.3802
No log 0.4375 14 1.8936 0.2835 1.8936 1.3761
No log 0.5 16 2.1268 0.2429 2.1268 1.4584
No log 0.5625 18 1.8557 0.3188 1.8557 1.3622
No log 0.625 20 1.6601 0.3731 1.6601 1.2885
No log 0.6875 22 1.7306 0.3429 1.7306 1.3155
No log 0.75 24 1.5468 0.3358 1.5468 1.2437
No log 0.8125 26 1.7820 0.3974 1.7820 1.3349
No log 0.875 28 1.8512 0.4 1.8512 1.3606
No log 0.9375 30 1.4347 0.4412 1.4347 1.1978
No log 1.0 32 1.2998 0.3607 1.2998 1.1401
No log 1.0625 34 1.2897 0.3968 1.2897 1.1356
No log 1.125 36 1.7132 0.4331 1.7132 1.3089
No log 1.1875 38 3.4653 0.1909 3.4653 1.8615
No log 1.25 40 5.3395 0.1201 5.3395 2.3107
No log 1.3125 42 4.9492 0.1026 4.9492 2.2247
No log 1.375 44 3.2379 0.2201 3.2379 1.7994
No log 1.4375 46 1.7068 0.3066 1.7068 1.3064
No log 1.5 48 1.4141 0.3902 1.4141 1.1892
No log 1.5625 50 1.4487 0.375 1.4487 1.2036
No log 1.625 52 1.6841 0.3704 1.6841 1.2977
No log 1.6875 54 1.9733 0.2361 1.9733 1.4048
No log 1.75 56 2.0015 0.2361 2.0015 1.4147
No log 1.8125 58 1.7158 0.3704 1.7158 1.3099
No log 1.875 60 1.6431 0.3824 1.6431 1.2818
No log 1.9375 62 1.8087 0.3624 1.8087 1.3449
No log 2.0 64 1.5831 0.4626 1.5831 1.2582
No log 2.0625 66 1.2271 0.5036 1.2271 1.1077
No log 2.125 68 1.1355 0.5468 1.1355 1.0656
No log 2.1875 70 1.3298 0.5576 1.3298 1.1532
No log 2.25 72 1.5869 0.5393 1.5869 1.2597
No log 2.3125 74 1.4905 0.5393 1.4905 1.2209
No log 2.375 76 1.1599 0.5513 1.1599 1.0770
No log 2.4375 78 1.1841 0.5616 1.1841 1.0882
No log 2.5 80 1.3944 0.5409 1.3944 1.1808
No log 2.5625 82 1.5933 0.5176 1.5933 1.2622
No log 2.625 84 1.1820 0.5733 1.1820 1.0872
No log 2.6875 86 0.8841 0.6232 0.8841 0.9403
No log 2.75 88 0.8585 0.6331 0.8585 0.9265
No log 2.8125 90 0.8866 0.6383 0.8866 0.9416
No log 2.875 92 0.9328 0.6490 0.9328 0.9658
No log 2.9375 94 0.8439 0.6522 0.8439 0.9187
No log 3.0 96 0.8654 0.6763 0.8654 0.9303
No log 3.0625 98 0.8612 0.6522 0.8612 0.9280
No log 3.125 100 0.9023 0.6525 0.9023 0.9499
No log 3.1875 102 0.9458 0.6389 0.9458 0.9725
No log 3.25 104 0.9046 0.6471 0.9046 0.9511
No log 3.3125 106 1.0856 0.5821 1.0856 1.0419
No log 3.375 108 1.1344 0.5373 1.1344 1.0651
No log 3.4375 110 0.9089 0.6519 0.9089 0.9534
No log 3.5 112 0.9552 0.625 0.9552 0.9774
No log 3.5625 114 1.5716 0.5464 1.5716 1.2536
No log 3.625 116 1.5746 0.5134 1.5746 1.2548
No log 3.6875 118 1.0371 0.6709 1.0371 1.0184
No log 3.75 120 0.8630 0.6571 0.8630 0.9290
No log 3.8125 122 1.0013 0.5865 1.0013 1.0007
No log 3.875 124 0.9384 0.6029 0.9384 0.9687
No log 3.9375 126 0.9896 0.6029 0.9896 0.9948
No log 4.0 128 1.4388 0.5581 1.4388 1.1995
No log 4.0625 130 1.5281 0.5030 1.5281 1.2362
No log 4.125 132 1.1640 0.5324 1.1640 1.0789
No log 4.1875 134 0.9670 0.6357 0.9670 0.9833
No log 4.25 136 1.0475 0.5827 1.0475 1.0235
No log 4.3125 138 1.0063 0.5714 1.0063 1.0032
No log 4.375 140 0.9900 0.6015 0.9900 0.9950
No log 4.4375 142 1.0878 0.5674 1.0878 1.0430
No log 4.5 144 1.0490 0.5899 1.0490 1.0242
No log 4.5625 146 0.9480 0.5970 0.9480 0.9737
No log 4.625 148 0.9208 0.5970 0.9208 0.9596
No log 4.6875 150 0.9651 0.5957 0.9651 0.9824
No log 4.75 152 1.0209 0.5972 1.0209 1.0104
No log 4.8125 154 0.9205 0.5674 0.9205 0.9594
No log 4.875 156 0.8176 0.7050 0.8176 0.9042
No log 4.9375 158 0.8539 0.6812 0.8539 0.9241
No log 5.0 160 0.9934 0.5816 0.9934 0.9967
No log 5.0625 162 0.8664 0.6761 0.8664 0.9308
No log 5.125 164 0.8135 0.7389 0.8135 0.9019
No log 5.1875 166 1.1133 0.6506 1.1133 1.0552
No log 5.25 168 1.1804 0.6429 1.1804 1.0865
No log 5.3125 170 0.9485 0.5986 0.9485 0.9739
No log 5.375 172 0.8604 0.6286 0.8604 0.9276
No log 5.4375 174 0.8217 0.6667 0.8217 0.9065
No log 5.5 176 0.9385 0.6040 0.9385 0.9688
No log 5.5625 178 1.1186 0.6065 1.1186 1.0576
No log 5.625 180 0.9813 0.6316 0.9813 0.9906
No log 5.6875 182 0.8136 0.6933 0.8136 0.9020
No log 5.75 184 0.7963 0.6667 0.7963 0.8923
No log 5.8125 186 0.8460 0.625 0.8460 0.9198
No log 5.875 188 0.9683 0.6358 0.9683 0.9840
No log 5.9375 190 0.9354 0.64 0.9354 0.9672
No log 6.0 192 0.8398 0.6154 0.8398 0.9164
No log 6.0625 194 0.8220 0.6475 0.8220 0.9067
No log 6.125 196 0.8237 0.6815 0.8237 0.9076
No log 6.1875 198 0.9477 0.6370 0.9477 0.9735
No log 6.25 200 0.9671 0.6418 0.9671 0.9834
No log 6.3125 202 0.8833 0.6519 0.8833 0.9398
No log 6.375 204 0.8086 0.6519 0.8086 0.8992
No log 6.4375 206 0.8265 0.7083 0.8265 0.9091
No log 6.5 208 0.8164 0.6715 0.8164 0.9035
No log 6.5625 210 0.8109 0.6765 0.8109 0.9005
No log 6.625 212 0.8239 0.7007 0.8239 0.9077
No log 6.6875 214 0.8369 0.6713 0.8369 0.9148
No log 6.75 216 1.0028 0.5974 1.0028 1.0014
No log 6.8125 218 1.0515 0.6303 1.0515 1.0254
No log 6.875 220 1.0059 0.6747 1.0059 1.0029
No log 6.9375 222 0.9307 0.6905 0.9307 0.9647
No log 7.0 224 0.9026 0.7006 0.9026 0.9501
No log 7.0625 226 1.0433 0.6784 1.0433 1.0214
No log 7.125 228 1.5435 0.5604 1.5435 1.2424
No log 7.1875 230 1.6726 0.5464 1.6726 1.2933
No log 7.25 232 1.2227 0.6047 1.2227 1.1057
No log 7.3125 234 0.8556 0.6525 0.8556 0.9250
No log 7.375 236 0.9206 0.6866 0.9206 0.9595
No log 7.4375 238 0.9002 0.6519 0.9002 0.9488
No log 7.5 240 0.8781 0.6324 0.8781 0.9371
No log 7.5625 242 1.0302 0.6438 1.0302 1.0150
No log 7.625 244 1.0737 0.6122 1.0737 1.0362
No log 7.6875 246 0.8985 0.6423 0.8985 0.9479
No log 7.75 248 0.8545 0.6165 0.8545 0.9244
No log 7.8125 250 0.8919 0.6466 0.8919 0.9444
No log 7.875 252 0.8386 0.6316 0.8386 0.9158
No log 7.9375 254 0.8503 0.6763 0.8503 0.9221
No log 8.0 256 0.9352 0.6667 0.9352 0.9670
No log 8.0625 258 0.9408 0.6338 0.9408 0.9699
No log 8.125 260 0.9057 0.6154 0.9057 0.9517
No log 8.1875 262 0.9234 0.6357 0.9234 0.9609
No log 8.25 264 0.9130 0.6357 0.9130 0.9555
No log 8.3125 266 0.8978 0.6107 0.8978 0.9475
No log 8.375 268 1.0475 0.5942 1.0475 1.0235
No log 8.4375 270 1.2015 0.5793 1.2015 1.0961
No log 8.5 272 1.2124 0.6118 1.2124 1.1011
No log 8.5625 274 0.9988 0.6429 0.9988 0.9994
No log 8.625 276 0.8476 0.6875 0.8476 0.9207
No log 8.6875 278 0.8149 0.7226 0.8149 0.9027
No log 8.75 280 0.7629 0.7114 0.7629 0.8734
No log 8.8125 282 0.7577 0.7050 0.7577 0.8705
No log 8.875 284 0.7752 0.7376 0.7752 0.8805
No log 8.9375 286 0.7734 0.7361 0.7734 0.8794
No log 9.0 288 0.7611 0.7092 0.7611 0.8724
No log 9.0625 290 0.7432 0.7211 0.7432 0.8621
No log 9.125 292 0.7396 0.7606 0.7396 0.8600
No log 9.1875 294 0.7532 0.7606 0.7532 0.8679
No log 9.25 296 0.7858 0.6906 0.7858 0.8865
No log 9.3125 298 0.8191 0.6906 0.8191 0.9051
No log 9.375 300 0.8095 0.6906 0.8095 0.8997
No log 9.4375 302 0.7973 0.6906 0.7973 0.8929
No log 9.5 304 0.8206 0.6714 0.8206 0.9059
No log 9.5625 306 0.8578 0.6 0.8578 0.9262
No log 9.625 308 0.8566 0.6232 0.8566 0.9255
No log 9.6875 310 0.8726 0.6232 0.8726 0.9341
No log 9.75 312 0.8420 0.6715 0.8420 0.9176
No log 9.8125 314 0.8978 0.6324 0.8978 0.9475
No log 9.875 316 0.8914 0.6324 0.8914 0.9441
No log 9.9375 318 0.8605 0.6471 0.8605 0.9276
No log 10.0 320 0.9023 0.6119 0.9023 0.9499
No log 10.0625 322 1.0206 0.5797 1.0206 1.0102
No log 10.125 324 1.0119 0.5496 1.0119 1.0060
No log 10.1875 326 0.9537 0.5736 0.9537 0.9766
No log 10.25 328 0.9113 0.5846 0.9113 0.9546
No log 10.3125 330 0.8423 0.7042 0.8423 0.9178
No log 10.375 332 0.8099 0.7347 0.8099 0.8999
No log 10.4375 334 0.8109 0.7320 0.8109 0.9005
No log 10.5 336 0.8484 0.6800 0.8484 0.9211
No log 10.5625 338 0.8507 0.6667 0.8507 0.9223
No log 10.625 340 0.8136 0.7034 0.8136 0.9020
No log 10.6875 342 0.7926 0.7361 0.7926 0.8903
No log 10.75 344 0.8070 0.7361 0.8070 0.8983
No log 10.8125 346 0.8441 0.6901 0.8441 0.9187
No log 10.875 348 0.9204 0.6573 0.9204 0.9594
No log 10.9375 350 0.9009 0.6712 0.9009 0.9492
No log 11.0 352 0.8177 0.7260 0.8177 0.9043
No log 11.0625 354 0.7892 0.7260 0.7892 0.8884
No log 11.125 356 0.7851 0.7260 0.7851 0.8861
No log 11.1875 358 0.7957 0.7172 0.7957 0.8920
No log 11.25 360 0.7993 0.7432 0.7993 0.8940
No log 11.3125 362 0.8260 0.7234 0.8260 0.9088
No log 11.375 364 0.8250 0.7092 0.8250 0.9083
No log 11.4375 366 0.8443 0.6957 0.8443 0.9188
No log 11.5 368 0.9286 0.6765 0.9286 0.9637
No log 11.5625 370 1.0320 0.5522 1.0320 1.0159
No log 11.625 372 0.9758 0.6370 0.9758 0.9878
No log 11.6875 374 0.8591 0.6765 0.8591 0.9269
No log 11.75 376 0.8417 0.75 0.8417 0.9174
No log 11.8125 378 0.8525 0.7134 0.8525 0.9233
No log 11.875 380 0.8769 0.6923 0.8769 0.9364
No log 11.9375 382 0.8667 0.6923 0.8667 0.9310
No log 12.0 384 0.8573 0.7059 0.8573 0.9259
No log 12.0625 386 0.8829 0.6767 0.8829 0.9396
No log 12.125 388 0.9462 0.6 0.9462 0.9727
No log 12.1875 390 1.0053 0.5846 1.0053 1.0027
No log 12.25 392 1.0332 0.5714 1.0332 1.0165
No log 12.3125 394 1.0362 0.5714 1.0362 1.0180
No log 12.375 396 0.9762 0.5926 0.9762 0.9880
No log 12.4375 398 0.8724 0.6667 0.8724 0.9340
No log 12.5 400 0.8606 0.6857 0.8606 0.9277
No log 12.5625 402 0.8222 0.7183 0.8222 0.9068
No log 12.625 404 0.7981 0.7273 0.7981 0.8934
No log 12.6875 406 0.8209 0.6812 0.8209 0.9060
No log 12.75 408 0.8508 0.6812 0.8508 0.9224
No log 12.8125 410 0.8751 0.6418 0.8751 0.9355
No log 12.875 412 0.8864 0.6260 0.8864 0.9415
No log 12.9375 414 0.9351 0.6406 0.9351 0.9670
No log 13.0 416 1.0540 0.6047 1.0540 1.0267
No log 13.0625 418 1.0070 0.6047 1.0070 1.0035
No log 13.125 420 0.8664 0.6912 0.8664 0.9308
No log 13.1875 422 0.7796 0.7050 0.7796 0.8829
No log 13.25 424 0.7769 0.7222 0.7769 0.8814
No log 13.3125 426 0.7678 0.7310 0.7678 0.8762
No log 13.375 428 0.7848 0.7324 0.7848 0.8859
No log 13.4375 430 0.7974 0.6857 0.7974 0.8930
No log 13.5 432 0.8268 0.6812 0.8268 0.9093
No log 13.5625 434 0.8682 0.6165 0.8682 0.9318
No log 13.625 436 0.9254 0.5581 0.9254 0.9620
No log 13.6875 438 0.9186 0.5581 0.9186 0.9584
No log 13.75 440 0.8549 0.6119 0.8549 0.9246
No log 13.8125 442 0.7971 0.7059 0.7971 0.8928
No log 13.875 444 0.7672 0.7153 0.7672 0.8759
No log 13.9375 446 0.7540 0.7133 0.7540 0.8683
No log 14.0 448 0.7714 0.6759 0.7714 0.8783
No log 14.0625 450 0.7955 0.6575 0.7955 0.8919
No log 14.125 452 0.8514 0.6577 0.8514 0.9227
No log 14.1875 454 0.8605 0.6577 0.8605 0.9277
No log 14.25 456 0.7903 0.6522 0.7903 0.8890
No log 14.3125 458 0.7735 0.6912 0.7735 0.8795
No log 14.375 460 0.7774 0.7007 0.7774 0.8817
No log 14.4375 462 0.8052 0.6260 0.8052 0.8973
No log 14.5 464 0.8509 0.5891 0.8509 0.9224
No log 14.5625 466 0.8533 0.5736 0.8533 0.9238
No log 14.625 468 0.8495 0.6667 0.8495 0.9217
No log 14.6875 470 0.8773 0.7068 0.8773 0.9366
No log 14.75 472 0.8853 0.7068 0.8853 0.9409
No log 14.8125 474 0.8661 0.6260 0.8661 0.9307
No log 14.875 476 0.8435 0.6324 0.8435 0.9184
No log 14.9375 478 0.8129 0.6423 0.8129 0.9016
No log 15.0 480 0.7918 0.6423 0.7918 0.8898
No log 15.0625 482 0.7601 0.6522 0.7601 0.8719
No log 15.125 484 0.7365 0.6763 0.7365 0.8582
No log 15.1875 486 0.7649 0.6522 0.7649 0.8746
No log 15.25 488 0.8315 0.6074 0.8315 0.9118
No log 15.3125 490 0.8799 0.5649 0.8799 0.9380
No log 15.375 492 0.8931 0.5909 0.8931 0.9450
No log 15.4375 494 0.8703 0.6466 0.8703 0.9329
No log 15.5 496 0.8568 0.6364 0.8568 0.9257
No log 15.5625 498 0.8437 0.6716 0.8437 0.9185
0.365 15.625 500 0.8250 0.6765 0.8250 0.9083
0.365 15.6875 502 0.8194 0.6861 0.8194 0.9052
0.365 15.75 504 0.8276 0.6861 0.8276 0.9097
0.365 15.8125 506 0.8310 0.6765 0.8310 0.9116
0.365 15.875 508 0.8309 0.6519 0.8309 0.9115
0.365 15.9375 510 0.8009 0.6714 0.8009 0.8949
0.365 16.0 512 0.7720 0.7042 0.7720 0.8786
0.365 16.0625 514 0.7709 0.6906 0.7709 0.8780
0.365 16.125 516 0.7595 0.7286 0.7595 0.8715
0.365 16.1875 518 0.7893 0.7338 0.7893 0.8884
0.365 16.25 520 0.8060 0.7518 0.8060 0.8978
0.365 16.3125 522 0.8243 0.7246 0.8243 0.9079
0.365 16.375 524 0.8648 0.6765 0.8648 0.9299
0.365 16.4375 526 0.9955 0.6617 0.9955 0.9977
0.365 16.5 528 1.0611 0.5312 1.0611 1.0301
0.365 16.5625 530 1.0330 0.6202 1.0330 1.0164
0.365 16.625 532 0.9557 0.5891 0.9557 0.9776

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k7_task1_organization

Finetuned
(4023)
this model