ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k7_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8957
  • Qwk: 0.6324
  • Mse: 0.8957
  • Rmse: 0.9464

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0625 2 7.0571 -0.0056 7.0571 2.6565
No log 0.125 4 4.1869 0.0855 4.1869 2.0462
No log 0.1875 6 3.1274 0.0727 3.1274 1.7684
No log 0.25 8 2.7210 0.0397 2.7210 1.6495
No log 0.3125 10 2.2940 0.1221 2.2940 1.5146
No log 0.375 12 2.1099 0.1429 2.1099 1.4526
No log 0.4375 14 1.7983 0.2222 1.7983 1.3410
No log 0.5 16 1.9490 0.1983 1.9490 1.3961
No log 0.5625 18 2.7411 0.0265 2.7411 1.6556
No log 0.625 20 2.6816 0.0263 2.6816 1.6376
No log 0.6875 22 2.1436 0.1194 2.1436 1.4641
No log 0.75 24 1.7778 0.3220 1.7778 1.3333
No log 0.8125 26 1.5181 0.2523 1.5181 1.2321
No log 0.875 28 1.4028 0.2342 1.4028 1.1844
No log 0.9375 30 1.3634 0.2182 1.3634 1.1676
No log 1.0 32 1.3915 0.3186 1.3915 1.1796
No log 1.0625 34 1.5888 0.3636 1.5888 1.2605
No log 1.125 36 2.0626 0.2174 2.0626 1.4362
No log 1.1875 38 2.6160 0.1294 2.6160 1.6174
No log 1.25 40 2.7273 0.1078 2.7273 1.6514
No log 1.3125 42 2.0471 0.28 2.0471 1.4308
No log 1.375 44 1.4968 0.4463 1.4968 1.2234
No log 1.4375 46 1.2936 0.4103 1.2936 1.1374
No log 1.5 48 1.2625 0.3793 1.2625 1.1236
No log 1.5625 50 1.3127 0.4390 1.3127 1.1457
No log 1.625 52 1.5009 0.4030 1.5009 1.2251
No log 1.6875 54 2.0010 0.3034 2.0010 1.4146
No log 1.75 56 2.1821 0.2368 2.1821 1.4772
No log 1.8125 58 1.7579 0.3404 1.7579 1.3258
No log 1.875 60 1.2781 0.4844 1.2781 1.1305
No log 1.9375 62 1.3908 0.4 1.3908 1.1793
No log 2.0 64 1.8917 0.1552 1.8917 1.3754
No log 2.0625 66 1.8123 0.2087 1.8123 1.3462
No log 2.125 68 1.6336 0.2632 1.6336 1.2781
No log 2.1875 70 1.4130 0.3158 1.4130 1.1887
No log 2.25 72 0.9914 0.5512 0.9914 0.9957
No log 2.3125 74 1.1042 0.5672 1.1042 1.0508
No log 2.375 76 1.3059 0.4895 1.3059 1.1428
No log 2.4375 78 1.1600 0.5414 1.1600 1.0770
No log 2.5 80 0.9559 0.6269 0.9559 0.9777
No log 2.5625 82 0.9783 0.5354 0.9783 0.9891
No log 2.625 84 1.0446 0.4516 1.0446 1.0221
No log 2.6875 86 0.9555 0.5538 0.9555 0.9775
No log 2.75 88 0.9034 0.5865 0.9034 0.9505
No log 2.8125 90 0.8985 0.6667 0.8985 0.9479
No log 2.875 92 0.9110 0.6763 0.9110 0.9545
No log 2.9375 94 0.9683 0.6571 0.9683 0.9840
No log 3.0 96 0.9325 0.6571 0.9325 0.9657
No log 3.0625 98 0.8505 0.6912 0.8505 0.9222
No log 3.125 100 0.9798 0.5312 0.9798 0.9898
No log 3.1875 102 1.1662 0.5 1.1662 1.0799
No log 3.25 104 1.0335 0.5312 1.0335 1.0166
No log 3.3125 106 0.9005 0.6418 0.9005 0.9489
No log 3.375 108 0.8434 0.6667 0.8434 0.9184
No log 3.4375 110 0.9343 0.6475 0.9343 0.9666
No log 3.5 112 1.1384 0.5564 1.1384 1.0669
No log 3.5625 114 1.5119 0.3485 1.5119 1.2296
No log 3.625 116 1.6010 0.3231 1.6010 1.2653
No log 3.6875 118 1.2425 0.5116 1.2425 1.1147
No log 3.75 120 1.0955 0.5649 1.0955 1.0466
No log 3.8125 122 1.2460 0.4590 1.2460 1.1162
No log 3.875 124 1.3615 0.3448 1.3615 1.1668
No log 3.9375 126 1.2157 0.4068 1.2157 1.1026
No log 4.0 128 1.1102 0.3932 1.1102 1.0537
No log 4.0625 130 1.1623 0.4390 1.1623 1.0781
No log 4.125 132 1.2271 0.5191 1.2271 1.1077
No log 4.1875 134 1.0940 0.5414 1.0940 1.0459
No log 4.25 136 0.8775 0.6617 0.8775 0.9368
No log 4.3125 138 0.8326 0.6957 0.8326 0.9125
No log 4.375 140 0.8368 0.7050 0.8368 0.9148
No log 4.4375 142 0.8298 0.7143 0.8298 0.9109
No log 4.5 144 0.8218 0.7222 0.8218 0.9066
No log 4.5625 146 0.8323 0.6803 0.8323 0.9123
No log 4.625 148 0.8729 0.6712 0.8729 0.9343
No log 4.6875 150 0.8926 0.625 0.8926 0.9448
No log 4.75 152 0.9441 0.6429 0.9441 0.9717
No log 4.8125 154 0.9482 0.6029 0.9482 0.9738
No log 4.875 156 0.9227 0.6429 0.9227 0.9606
No log 4.9375 158 0.9654 0.6618 0.9654 0.9825
No log 5.0 160 1.0039 0.6519 1.0039 1.0020
No log 5.0625 162 0.9877 0.6519 0.9877 0.9938
No log 5.125 164 0.9827 0.6519 0.9827 0.9913
No log 5.1875 166 0.9766 0.5426 0.9766 0.9882
No log 5.25 168 0.9930 0.528 0.9930 0.9965
No log 5.3125 170 1.0250 0.528 1.0250 1.0124
No log 5.375 172 1.0519 0.5 1.0519 1.0256
No log 5.4375 174 1.0336 0.5469 1.0336 1.0167
No log 5.5 176 0.9867 0.5669 0.9867 0.9933
No log 5.5625 178 0.9904 0.5909 0.9904 0.9952
No log 5.625 180 0.9873 0.6765 0.9873 0.9936
No log 5.6875 182 0.9383 0.6861 0.9383 0.9687
No log 5.75 184 0.9445 0.6087 0.9445 0.9719
No log 5.8125 186 0.9981 0.6197 0.9981 0.9990
No log 5.875 188 1.0376 0.5986 1.0376 1.0186
No log 5.9375 190 1.0073 0.5915 1.0073 1.0036
No log 6.0 192 0.9523 0.6241 0.9523 0.9759
No log 6.0625 194 0.9070 0.6143 0.9070 0.9524
No log 6.125 196 0.8969 0.6471 0.8969 0.9470
No log 6.1875 198 0.8812 0.6667 0.8812 0.9387
No log 6.25 200 0.8723 0.6667 0.8723 0.9339
No log 6.3125 202 0.8841 0.6569 0.8841 0.9403
No log 6.375 204 0.8932 0.6324 0.8932 0.9451
No log 6.4375 206 0.9972 0.6522 0.9972 0.9986
No log 6.5 208 1.0337 0.6423 1.0337 1.0167
No log 6.5625 210 0.9837 0.6131 0.9837 0.9918
No log 6.625 212 0.9105 0.6618 0.9105 0.9542
No log 6.6875 214 0.9091 0.6074 0.9091 0.9535
No log 6.75 216 0.9079 0.6423 0.9079 0.9528
No log 6.8125 218 0.9334 0.6944 0.9334 0.9661
No log 6.875 220 1.0128 0.6154 1.0128 1.0064
No log 6.9375 222 1.0108 0.6763 1.0108 1.0054
No log 7.0 224 0.9658 0.6370 0.9658 0.9828
No log 7.0625 226 0.9927 0.6015 0.9927 0.9963
No log 7.125 228 1.0205 0.5758 1.0205 1.0102
No log 7.1875 230 1.0504 0.5469 1.0504 1.0249
No log 7.25 232 1.0227 0.5865 1.0227 1.0113
No log 7.3125 234 1.0018 0.6324 1.0018 1.0009
No log 7.375 236 1.0304 0.6331 1.0304 1.0151
No log 7.4375 238 1.0102 0.6324 1.0102 1.0051
No log 7.5 240 0.9735 0.6015 0.9735 0.9867
No log 7.5625 242 0.9843 0.5954 0.9843 0.9921
No log 7.625 244 1.0111 0.5496 1.0111 1.0055
No log 7.6875 246 0.9597 0.5846 0.9597 0.9797
No log 7.75 248 0.9174 0.6212 0.9174 0.9578
No log 7.8125 250 0.8826 0.6667 0.8826 0.9395
No log 7.875 252 0.8882 0.6176 0.8882 0.9424
No log 7.9375 254 0.9207 0.6324 0.9207 0.9595
No log 8.0 256 0.9864 0.6522 0.9864 0.9932
No log 8.0625 258 0.9770 0.6222 0.9770 0.9885
No log 8.125 260 0.9204 0.6286 0.9204 0.9594
No log 8.1875 262 0.8681 0.6806 0.8681 0.9317
No log 8.25 264 0.8623 0.6806 0.8623 0.9286
No log 8.3125 266 0.8763 0.6901 0.8763 0.9361
No log 8.375 268 0.9054 0.6232 0.9054 0.9516
No log 8.4375 270 0.9443 0.5865 0.9443 0.9717
No log 8.5 272 0.9094 0.6370 0.9094 0.9536
No log 8.5625 274 0.8633 0.6165 0.8633 0.9291
No log 8.625 276 0.8344 0.6714 0.8344 0.9135
No log 8.6875 278 0.8407 0.6763 0.8407 0.9169
No log 8.75 280 0.9547 0.6906 0.9547 0.9771
No log 8.8125 282 1.0694 0.6074 1.0694 1.0341
No log 8.875 284 1.0167 0.7007 1.0167 1.0083
No log 8.9375 286 0.9129 0.6906 0.9129 0.9554
No log 9.0 288 0.8609 0.6418 0.8609 0.9278
No log 9.0625 290 0.8856 0.6324 0.8856 0.9410
No log 9.125 292 0.8904 0.5909 0.8904 0.9436
No log 9.1875 294 0.9061 0.6418 0.9061 0.9519
No log 9.25 296 0.9199 0.6000 0.9199 0.9591
No log 9.3125 298 0.9194 0.6107 0.9194 0.9588
No log 9.375 300 0.9070 0.6567 0.9070 0.9524
No log 9.4375 302 0.8741 0.6861 0.8741 0.9349
No log 9.5 304 0.8817 0.6853 0.8817 0.9390
No log 9.5625 306 0.9396 0.6479 0.9396 0.9693
No log 9.625 308 0.9319 0.6479 0.9319 0.9653
No log 9.6875 310 0.9717 0.6479 0.9717 0.9857
No log 9.75 312 0.9702 0.6479 0.9702 0.9850
No log 9.8125 314 0.9608 0.6906 0.9608 0.9802
No log 9.875 316 0.9337 0.6857 0.9337 0.9663
No log 9.9375 318 0.8569 0.7143 0.8569 0.9257
No log 10.0 320 0.7960 0.7234 0.7960 0.8922
No log 10.0625 322 0.7861 0.7260 0.7861 0.8866
No log 10.125 324 0.7746 0.7211 0.7746 0.8801
No log 10.1875 326 0.7080 0.7532 0.7080 0.8414
No log 10.25 328 0.7141 0.7355 0.7141 0.8451
No log 10.3125 330 0.7980 0.6755 0.7980 0.8933
No log 10.375 332 0.8313 0.6486 0.8313 0.9118
No log 10.4375 334 0.8323 0.6667 0.8323 0.9123
No log 10.5 336 0.8547 0.6944 0.8547 0.9245
No log 10.5625 338 0.9272 0.7042 0.9272 0.9629
No log 10.625 340 0.9670 0.6377 0.9670 0.9834
No log 10.6875 342 0.9559 0.6763 0.9559 0.9777
No log 10.75 344 0.9682 0.6620 0.9682 0.9840
No log 10.8125 346 0.9661 0.6434 0.9661 0.9829
No log 10.875 348 0.9611 0.625 0.9611 0.9804
No log 10.9375 350 0.9641 0.6154 0.9641 0.9819
No log 11.0 352 0.9235 0.6667 0.9235 0.9610
No log 11.0625 354 0.9520 0.6187 0.9520 0.9757
No log 11.125 356 1.0202 0.6015 1.0202 1.0100
No log 11.1875 358 1.1452 0.5909 1.1452 1.0701
No log 11.25 360 1.1752 0.5630 1.1752 1.0841
No log 11.3125 362 1.2132 0.5109 1.2132 1.1014
No log 11.375 364 1.1240 0.5401 1.1240 1.0602
No log 11.4375 366 1.0459 0.6331 1.0459 1.0227
No log 11.5 368 0.9391 0.6957 0.9391 0.9691
No log 11.5625 370 0.8625 0.6715 0.8625 0.9287
No log 11.625 372 0.8499 0.6715 0.8499 0.9219
No log 11.6875 374 0.9040 0.6763 0.9040 0.9508
No log 11.75 376 0.9545 0.6667 0.9545 0.9770
No log 11.8125 378 0.9398 0.6715 0.9398 0.9694
No log 11.875 380 0.9205 0.6 0.9205 0.9594
No log 11.9375 382 0.9697 0.5312 0.9697 0.9847
No log 12.0 384 0.9505 0.5581 0.9505 0.9749
No log 12.0625 386 0.8975 0.6061 0.8975 0.9474
No log 12.125 388 0.8642 0.6715 0.8642 0.9296
No log 12.1875 390 0.8147 0.6957 0.8147 0.9026
No log 12.25 392 0.7778 0.7042 0.7778 0.8819
No log 12.3125 394 0.8072 0.6207 0.8072 0.8984
No log 12.375 396 0.8148 0.6207 0.8148 0.9026
No log 12.4375 398 0.7787 0.7114 0.7787 0.8825
No log 12.5 400 0.8078 0.6993 0.8078 0.8988
No log 12.5625 402 0.8692 0.7 0.8692 0.9323
No log 12.625 404 0.9193 0.6906 0.9193 0.9588
No log 12.6875 406 0.9660 0.6154 0.9660 0.9829
No log 12.75 408 1.0122 0.5714 1.0122 1.0061
No log 12.8125 410 1.0280 0.5082 1.0280 1.0139
No log 12.875 412 1.0131 0.5354 1.0131 1.0066
No log 12.9375 414 0.9959 0.5426 0.9959 0.9979
No log 13.0 416 0.9468 0.5970 0.9468 0.9730
No log 13.0625 418 0.8852 0.7042 0.8852 0.9408
No log 13.125 420 0.8586 0.7042 0.8586 0.9266
No log 13.1875 422 0.8432 0.6619 0.8432 0.9183
No log 13.25 424 0.8445 0.6857 0.8445 0.9190
No log 13.3125 426 0.8437 0.6857 0.8437 0.9185
No log 13.375 428 0.8513 0.6667 0.8513 0.9227
No log 13.4375 430 0.8790 0.6861 0.8790 0.9376
No log 13.5 432 0.9235 0.6222 0.9235 0.9610
No log 13.5625 434 0.9358 0.6222 0.9358 0.9674
No log 13.625 436 0.9140 0.6377 0.9140 0.9560
No log 13.6875 438 0.9137 0.6222 0.9137 0.9559
No log 13.75 440 0.9194 0.6222 0.9194 0.9589
No log 13.8125 442 0.9072 0.6466 0.9072 0.9525
No log 13.875 444 0.9392 0.5669 0.9392 0.9691
No log 13.9375 446 0.9616 0.5669 0.9616 0.9806
No log 14.0 448 0.9653 0.5669 0.9653 0.9825
No log 14.0625 450 0.9349 0.6 0.9350 0.9669
No log 14.125 452 0.9016 0.6260 0.9016 0.9495
No log 14.1875 454 0.8689 0.6815 0.8689 0.9322
No log 14.25 456 0.8694 0.6716 0.8694 0.9324
No log 14.3125 458 0.9048 0.6569 0.9048 0.9512
No log 14.375 460 0.9217 0.6522 0.9217 0.9601
No log 14.4375 462 0.9091 0.6619 0.9091 0.9535
No log 14.5 464 0.8777 0.6522 0.8777 0.9369
No log 14.5625 466 0.8345 0.7007 0.8345 0.9135
No log 14.625 468 0.8248 0.6912 0.8248 0.9082
No log 14.6875 470 0.8377 0.6912 0.8377 0.9152
No log 14.75 472 0.8634 0.7007 0.8634 0.9292
No log 14.8125 474 0.8759 0.6912 0.8759 0.9359
No log 14.875 476 0.9137 0.6569 0.9137 0.9559
No log 14.9375 478 0.9444 0.6618 0.9444 0.9718
No log 15.0 480 0.9330 0.6567 0.9330 0.9659
No log 15.0625 482 0.9001 0.6569 0.9001 0.9488
No log 15.125 484 0.8499 0.7092 0.8499 0.9219
No log 15.1875 486 0.8434 0.7413 0.8434 0.9184
No log 15.25 488 0.8756 0.6993 0.8756 0.9358
No log 15.3125 490 0.8992 0.6713 0.8992 0.9482
No log 15.375 492 0.8619 0.6809 0.8619 0.9284
No log 15.4375 494 0.8374 0.7 0.8374 0.9151
No log 15.5 496 0.8529 0.6906 0.8529 0.9235
No log 15.5625 498 0.8898 0.6519 0.8898 0.9433
0.3979 15.625 500 0.9793 0.6667 0.9793 0.9896
0.3979 15.6875 502 1.0351 0.6861 1.0351 1.0174
0.3979 15.75 504 1.0471 0.6715 1.0471 1.0233
0.3979 15.8125 506 1.0107 0.6165 1.0107 1.0053
0.3979 15.875 508 0.9593 0.6466 0.9593 0.9794
0.3979 15.9375 510 0.9159 0.6519 0.9159 0.9570
0.3979 16.0 512 0.8972 0.6618 0.8972 0.9472
0.3979 16.0625 514 0.8957 0.6324 0.8957 0.9464

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run3_AugV5_k7_task1_organization

Finetuned
(4023)
this model