ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k6_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8483
  • Qwk: 0.6718
  • Mse: 0.8483
  • Rmse: 0.9211

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 6.7723 0.0114 6.7723 2.6024
No log 0.1379 4 4.3788 0.0080 4.3788 2.0925
No log 0.2069 6 3.0747 0.0119 3.0747 1.7535
No log 0.2759 8 2.8001 0.0870 2.8001 1.6733
No log 0.3448 10 2.4049 0.1240 2.4049 1.5508
No log 0.4138 12 1.8839 0.0536 1.8839 1.3725
No log 0.4828 14 1.8022 0.0769 1.8022 1.3425
No log 0.5517 16 1.9023 0.1143 1.9023 1.3792
No log 0.6207 18 1.9104 0.1121 1.9104 1.3822
No log 0.6897 20 2.1740 0.1739 2.1740 1.4745
No log 0.7586 22 2.3018 0.0694 2.3018 1.5172
No log 0.8276 24 2.4737 0.0274 2.4737 1.5728
No log 0.8966 26 2.8234 0.0129 2.8234 1.6803
No log 0.9655 28 3.1126 0.0252 3.1126 1.7643
No log 1.0345 30 2.5216 0.1517 2.5216 1.5880
No log 1.1034 32 1.5946 0.3089 1.5946 1.2628
No log 1.1724 34 1.7266 0.2243 1.7266 1.3140
No log 1.2414 36 2.1441 -0.0357 2.1441 1.4643
No log 1.3103 38 2.3955 0.0455 2.3955 1.5478
No log 1.3793 40 2.3882 0.0420 2.3882 1.5454
No log 1.4483 42 2.3090 0.0414 2.3090 1.5195
No log 1.5172 44 2.1648 0.1806 2.1648 1.4713
No log 1.5862 46 2.0161 0.2714 2.0161 1.4199
No log 1.6552 48 1.9778 0.3262 1.9778 1.4064
No log 1.7241 50 2.1093 0.3221 2.1093 1.4524
No log 1.7931 52 2.0527 0.3514 2.0527 1.4327
No log 1.8621 54 2.2170 0.2914 2.2170 1.4890
No log 1.9310 56 2.6891 0.2584 2.6891 1.6398
No log 2.0 58 2.6492 0.2712 2.6492 1.6276
No log 2.0690 60 2.0788 0.3268 2.0788 1.4418
No log 2.1379 62 1.6823 0.3150 1.6823 1.2970
No log 2.2069 64 1.4944 0.3103 1.4944 1.2225
No log 2.2759 66 1.5165 0.2523 1.5165 1.2315
No log 2.3448 68 1.5717 0.2478 1.5717 1.2537
No log 2.4138 70 1.5704 0.3802 1.5704 1.2532
No log 2.4828 72 1.7185 0.3759 1.7185 1.3109
No log 2.5517 74 2.2767 0.3049 2.2767 1.5089
No log 2.6207 76 2.6549 0.2472 2.6549 1.6294
No log 2.6897 78 2.5129 0.2825 2.5129 1.5852
No log 2.7586 80 2.0934 0.2953 2.0934 1.4469
No log 2.8276 82 1.7813 0.2979 1.7813 1.3347
No log 2.8966 84 1.6003 0.3538 1.6003 1.2650
No log 2.9655 86 1.4245 0.4409 1.4245 1.1935
No log 3.0345 88 1.5302 0.4203 1.5302 1.2370
No log 3.1034 90 1.6808 0.3636 1.6808 1.2965
No log 3.1724 92 1.3874 0.4593 1.3874 1.1779
No log 3.2414 94 1.3144 0.4928 1.3144 1.1465
No log 3.3103 96 1.3391 0.5263 1.3391 1.1572
No log 3.3793 98 1.5351 0.4795 1.5351 1.2390
No log 3.4483 100 1.6463 0.4889 1.6463 1.2831
No log 3.5172 102 2.1102 0.4352 2.1102 1.4526
No log 3.5862 104 1.9252 0.4792 1.9252 1.3875
No log 3.6552 106 1.2945 0.5399 1.2945 1.1378
No log 3.7241 108 1.1420 0.5185 1.1420 1.0686
No log 3.7931 110 1.0778 0.5672 1.0778 1.0382
No log 3.8621 112 0.9869 0.5985 0.9869 0.9934
No log 3.9310 114 0.9578 0.6331 0.9578 0.9787
No log 4.0 116 0.9582 0.6176 0.9582 0.9789
No log 4.0690 118 0.9303 0.6377 0.9303 0.9645
No log 4.1379 120 0.9032 0.6475 0.9032 0.9504
No log 4.2069 122 0.9014 0.6197 0.9014 0.9494
No log 4.2759 124 0.9328 0.6207 0.9328 0.9658
No log 4.3448 126 0.8806 0.6164 0.8806 0.9384
No log 4.4138 128 0.8821 0.7702 0.8821 0.9392
No log 4.4828 130 0.9709 0.7093 0.9709 0.9853
No log 4.5517 132 0.8532 0.7425 0.8532 0.9237
No log 4.6207 134 0.8366 0.72 0.8366 0.9146
No log 4.6897 136 0.8796 0.6438 0.8796 0.9379
No log 4.7586 138 0.8015 0.7432 0.8015 0.8953
No log 4.8276 140 0.8004 0.7821 0.8004 0.8947
No log 4.8966 142 0.8916 0.7251 0.8916 0.9443
No log 4.9655 144 0.8324 0.7647 0.8324 0.9123
No log 5.0345 146 0.8231 0.7067 0.8231 0.9072
No log 5.1034 148 1.0085 0.6154 1.0085 1.0043
No log 5.1724 150 0.9727 0.6241 0.9727 0.9863
No log 5.2414 152 0.8394 0.7310 0.8394 0.9162
No log 5.3103 154 0.8287 0.7320 0.8287 0.9103
No log 5.3793 156 0.8275 0.7347 0.8275 0.9097
No log 5.4483 158 0.8254 0.7260 0.8254 0.9085
No log 5.5172 160 0.8142 0.7248 0.8142 0.9023
No log 5.5862 162 0.8274 0.7222 0.8274 0.9096
No log 5.6552 164 0.9621 0.6232 0.9621 0.9809
No log 5.7241 166 1.0258 0.5672 1.0258 1.0128
No log 5.7931 168 1.0095 0.6383 1.0095 1.0048
No log 5.8621 170 1.0869 0.5634 1.0869 1.0425
No log 5.9310 172 1.2355 0.5369 1.2355 1.1115
No log 6.0 174 1.1968 0.5315 1.1968 1.0940
No log 6.0690 176 1.0437 0.5674 1.0437 1.0216
No log 6.1379 178 0.9393 0.6763 0.9393 0.9692
No log 6.2069 180 0.9186 0.6667 0.9186 0.9584
No log 6.2759 182 0.8658 0.6857 0.8658 0.9305
No log 6.3448 184 0.8177 0.7432 0.8177 0.9043
No log 6.4138 186 0.9770 0.65 0.9770 0.9884
No log 6.4828 188 1.0614 0.6548 1.0614 1.0302
No log 6.5517 190 0.8774 0.7097 0.8774 0.9367
No log 6.6207 192 0.8072 0.7432 0.8072 0.8985
No log 6.6897 194 0.8058 0.7417 0.8058 0.8977
No log 6.7586 196 0.8393 0.7075 0.8393 0.9162
No log 6.8276 198 0.8229 0.7516 0.8229 0.9071
No log 6.8966 200 0.8318 0.7013 0.8318 0.9120
No log 6.9655 202 0.8469 0.6667 0.8469 0.9203
No log 7.0345 204 0.8621 0.6533 0.8621 0.9285
No log 7.1034 206 0.8190 0.7075 0.8190 0.9050
No log 7.1724 208 0.8243 0.7075 0.8243 0.9079
No log 7.2414 210 0.8116 0.6939 0.8116 0.9009
No log 7.3103 212 0.8430 0.6928 0.8430 0.9181
No log 7.3793 214 0.8352 0.7067 0.8352 0.9139
No log 7.4483 216 0.8393 0.7190 0.8393 0.9161
No log 7.5172 218 0.8409 0.7051 0.8409 0.9170
No log 7.5862 220 0.8360 0.6980 0.8360 0.9144
No log 7.6552 222 0.8496 0.6993 0.8496 0.9217
No log 7.7241 224 0.8419 0.7067 0.8419 0.9175
No log 7.7931 226 0.9644 0.6494 0.9644 0.9820
No log 7.8621 228 0.9952 0.6405 0.9952 0.9976
No log 7.9310 230 0.8893 0.7190 0.8893 0.9430
No log 8.0 232 0.8264 0.7190 0.8264 0.9091
No log 8.0690 234 0.8267 0.7083 0.8267 0.9092
No log 8.1379 236 0.8336 0.6897 0.8336 0.9130
No log 8.2069 238 0.8368 0.7114 0.8368 0.9148
No log 8.2759 240 0.8177 0.6759 0.8177 0.9042
No log 8.3448 242 0.8062 0.6897 0.8062 0.8979
No log 8.4138 244 0.8612 0.6471 0.8612 0.9280
No log 8.4828 246 0.8784 0.6912 0.8784 0.9372
No log 8.5517 248 0.8210 0.6950 0.8210 0.9061
No log 8.6207 250 0.7879 0.7662 0.7879 0.8876
No log 8.6897 252 0.7998 0.7226 0.7998 0.8943
No log 8.7586 254 0.7885 0.7662 0.7885 0.8880
No log 8.8276 256 0.7868 0.7517 0.7868 0.8870
No log 8.8966 258 0.7699 0.7895 0.7699 0.8774
No log 8.9655 260 0.7493 0.8 0.7493 0.8656
No log 9.0345 262 0.7664 0.7342 0.7664 0.8755
No log 9.1034 264 0.8193 0.7205 0.8193 0.9052
No log 9.1724 266 0.8427 0.7160 0.8427 0.9180
No log 9.2414 268 0.8112 0.7081 0.8112 0.9007
No log 9.3103 270 0.7496 0.7792 0.7496 0.8658
No log 9.3793 272 0.8805 0.7194 0.8805 0.9384
No log 9.4483 274 0.9266 0.6815 0.9266 0.9626
No log 9.5172 276 0.8831 0.7299 0.8831 0.9398
No log 9.5862 278 0.8581 0.7286 0.8581 0.9263
No log 9.6552 280 0.8654 0.6667 0.8654 0.9303
No log 9.7241 282 0.8774 0.6619 0.8774 0.9367
No log 9.7931 284 0.8679 0.6429 0.8679 0.9316
No log 9.8621 286 0.8751 0.6667 0.8751 0.9355
No log 9.9310 288 0.8555 0.7237 0.8555 0.9249
No log 10.0 290 0.8304 0.7179 0.8304 0.9113
No log 10.0690 292 0.8124 0.7285 0.8124 0.9013
No log 10.1379 294 0.8436 0.7417 0.8436 0.9185
No log 10.2069 296 0.8612 0.7248 0.8612 0.9280
No log 10.2759 298 0.8653 0.6892 0.8653 0.9302
No log 10.3448 300 0.8483 0.7075 0.8483 0.9211
No log 10.4138 302 0.8534 0.7333 0.8534 0.9238
No log 10.4828 304 0.9057 0.6099 0.9057 0.9517
No log 10.5517 306 0.8929 0.6522 0.8929 0.9449
No log 10.6207 308 0.8858 0.6377 0.8858 0.9411
No log 10.6897 310 0.8592 0.6968 0.8592 0.9269
No log 10.7586 312 0.7991 0.7211 0.7991 0.8939
No log 10.8276 314 0.7755 0.7297 0.7755 0.8806
No log 10.8966 316 0.7619 0.7226 0.7619 0.8728
No log 10.9655 318 0.7565 0.7134 0.7565 0.8698
No log 11.0345 320 0.7747 0.7432 0.7747 0.8802
No log 11.1034 322 0.7946 0.7 0.7946 0.8914
No log 11.1724 324 0.7979 0.7237 0.7979 0.8933
No log 11.2414 326 0.7876 0.7194 0.7876 0.8875
No log 11.3103 328 0.8148 0.7111 0.8148 0.9027
No log 11.3793 330 0.8027 0.7338 0.8027 0.8959
No log 11.4483 332 0.7945 0.7194 0.7945 0.8913
No log 11.5172 334 0.8200 0.6757 0.8200 0.9056
No log 11.5862 336 0.8345 0.7134 0.8345 0.9135
No log 11.6552 338 0.7940 0.6892 0.7940 0.8911
No log 11.7241 340 0.7805 0.7123 0.7805 0.8835
No log 11.7931 342 0.8052 0.7261 0.8052 0.8973
No log 11.8621 344 0.8107 0.7215 0.8107 0.9004
No log 11.9310 346 0.7770 0.7702 0.7770 0.8815
No log 12.0 348 0.7635 0.7682 0.7635 0.8738
No log 12.0690 350 0.7688 0.7742 0.7688 0.8768
No log 12.1379 352 0.7809 0.7636 0.7809 0.8837
No log 12.2069 354 0.8307 0.7152 0.8307 0.9114
No log 12.2759 356 0.8107 0.7362 0.8107 0.9004
No log 12.3448 358 0.7659 0.7662 0.7659 0.8751
No log 12.4138 360 0.8093 0.7286 0.8093 0.8996
No log 12.4828 362 0.8498 0.7007 0.8498 0.9219
No log 12.5517 364 0.8165 0.6861 0.8165 0.9036
No log 12.6207 366 0.8209 0.6901 0.8209 0.9060
No log 12.6897 368 0.8790 0.6933 0.8790 0.9375
No log 12.7586 370 0.9224 0.6316 0.9224 0.9604
No log 12.8276 372 0.9028 0.6710 0.9028 0.9502
No log 12.8966 374 0.8410 0.7013 0.8410 0.9171
No log 12.9655 376 0.7669 0.7285 0.7669 0.8757
No log 13.0345 378 0.7756 0.7226 0.7756 0.8807
No log 13.1034 380 0.8374 0.7051 0.8374 0.9151
No log 13.1724 382 0.8456 0.7097 0.8456 0.9196
No log 13.2414 384 0.8123 0.6906 0.8123 0.9013
No log 13.3103 386 0.7945 0.6912 0.7945 0.8914
No log 13.3793 388 0.8147 0.6815 0.8147 0.9026
No log 13.4483 390 0.8408 0.6815 0.8408 0.9170
No log 13.5172 392 0.8766 0.6370 0.8766 0.9363
No log 13.5862 394 0.8575 0.6715 0.8575 0.9260
No log 13.6552 396 0.8397 0.6861 0.8397 0.9163
No log 13.7241 398 0.8462 0.6861 0.8462 0.9199
No log 13.7931 400 0.8760 0.6763 0.8760 0.9360
No log 13.8621 402 0.8882 0.6765 0.8882 0.9424
No log 13.9310 404 0.8657 0.6765 0.8657 0.9304
No log 14.0 406 0.8450 0.7222 0.8450 0.9192
No log 14.0690 408 0.9076 0.6711 0.9076 0.9527
No log 14.1379 410 0.9016 0.6711 0.9016 0.9495
No log 14.2069 412 0.8043 0.7260 0.8043 0.8968
No log 14.2759 414 0.8111 0.6763 0.8111 0.9006
No log 14.3448 416 0.9625 0.5797 0.9625 0.9811
No log 14.4138 418 1.0577 0.5957 1.0577 1.0284
No log 14.4828 420 1.0040 0.5564 1.0040 1.0020
No log 14.5517 422 0.8938 0.5802 0.8938 0.9454
No log 14.6207 424 0.8664 0.7111 0.8664 0.9308
No log 14.6897 426 0.8727 0.7246 0.8727 0.9342
No log 14.7586 428 0.8527 0.7246 0.8527 0.9234
No log 14.8276 430 0.8184 0.7376 0.8184 0.9047
No log 14.8966 432 0.7882 0.75 0.7882 0.8878
No log 14.9655 434 0.7889 0.75 0.7889 0.8882
No log 15.0345 436 0.7793 0.75 0.7793 0.8828
No log 15.1034 438 0.7823 0.75 0.7823 0.8845
No log 15.1724 440 0.7834 0.7413 0.7834 0.8851
No log 15.2414 442 0.8071 0.7050 0.8071 0.8984
No log 15.3103 444 0.8376 0.6667 0.8376 0.9152
No log 15.3793 446 0.8229 0.7050 0.8229 0.9071
No log 15.4483 448 0.7943 0.6912 0.7943 0.8912
No log 15.5172 450 0.7647 0.7246 0.7647 0.8745
No log 15.5862 452 0.7364 0.7183 0.7364 0.8582
No log 15.6552 454 0.7292 0.7183 0.7292 0.8539
No log 15.7241 456 0.7240 0.7376 0.7240 0.8509
No log 15.7931 458 0.7497 0.7286 0.7497 0.8659
No log 15.8621 460 0.7981 0.7050 0.7981 0.8933
No log 15.9310 462 0.8528 0.6515 0.8528 0.9235
No log 16.0 464 0.9095 0.5954 0.9095 0.9537
No log 16.0690 466 0.9401 0.6015 0.9401 0.9696
No log 16.1379 468 0.9482 0.5802 0.9482 0.9738
No log 16.2069 470 0.9392 0.5909 0.9392 0.9691
No log 16.2759 472 0.8792 0.6015 0.8792 0.9376
No log 16.3448 474 0.8421 0.6277 0.8421 0.9177
No log 16.4138 476 0.8432 0.6667 0.8432 0.9183
No log 16.4828 478 0.8312 0.6763 0.8312 0.9117
No log 16.5517 480 0.7876 0.6963 0.7876 0.8875
No log 16.6207 482 0.7741 0.6767 0.7741 0.8798
No log 16.6897 484 0.7777 0.6963 0.7777 0.8818
No log 16.7586 486 0.7795 0.7015 0.7795 0.8829
No log 16.8276 488 0.7839 0.7194 0.7839 0.8854
No log 16.8966 490 0.7881 0.6715 0.7881 0.8877
No log 16.9655 492 0.7677 0.7259 0.7677 0.8762
No log 17.0345 494 0.7955 0.6963 0.7955 0.8919
No log 17.1034 496 0.7972 0.7068 0.7972 0.8929
No log 17.1724 498 0.7601 0.7391 0.7601 0.8718
0.4287 17.2414 500 0.7591 0.7222 0.7591 0.8713
0.4287 17.3103 502 0.8021 0.7355 0.8021 0.8956
0.4287 17.3793 504 0.7853 0.7355 0.7853 0.8862
0.4287 17.4483 506 0.7383 0.7417 0.7383 0.8593
0.4287 17.5172 508 0.7293 0.7552 0.7293 0.8540
0.4287 17.5862 510 0.7941 0.7 0.7941 0.8911
0.4287 17.6552 512 0.8333 0.6763 0.8333 0.9128
0.4287 17.7241 514 0.8001 0.7246 0.8001 0.8945
0.4287 17.7931 516 0.7842 0.7050 0.7842 0.8855
0.4287 17.8621 518 0.7786 0.7050 0.7786 0.8824
0.4287 17.9310 520 0.7740 0.7397 0.7740 0.8798
0.4287 18.0 522 0.7450 0.7397 0.7450 0.8632
0.4287 18.0690 524 0.7268 0.7778 0.7268 0.8525
0.4287 18.1379 526 0.7411 0.7660 0.7411 0.8609
0.4287 18.2069 528 0.7659 0.7571 0.7659 0.8752
0.4287 18.2759 530 0.8017 0.6866 0.8017 0.8954
0.4287 18.3448 532 0.8402 0.6866 0.8402 0.9166
0.4287 18.4138 534 0.8684 0.6357 0.8684 0.9319
0.4287 18.4828 536 0.8662 0.6718 0.8662 0.9307
0.4287 18.5517 538 0.8483 0.6718 0.8483 0.9211

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k6_task1_organization

Finetuned
(4023)
this model