ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k18_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7839
  • Qwk: 0.7211
  • Mse: 0.7839
  • Rmse: 0.8854

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0233 2 6.9214 0.0057 6.9214 2.6309
No log 0.0465 4 4.7085 0.0576 4.7085 2.1699
No log 0.0698 6 4.4413 -0.0773 4.4413 2.1074
No log 0.0930 8 3.7931 -0.0305 3.7931 1.9476
No log 0.1163 10 2.5133 0.1277 2.5133 1.5853
No log 0.1395 12 2.1914 0.1583 2.1914 1.4803
No log 0.1628 14 2.3688 0.0986 2.3688 1.5391
No log 0.1860 16 2.2377 0.1460 2.2377 1.4959
No log 0.2093 18 2.3143 0.1127 2.3143 1.5213
No log 0.2326 20 2.3296 0.0694 2.3296 1.5263
No log 0.2558 22 2.1241 0.2353 2.1241 1.4574
No log 0.2791 24 1.8677 0.3252 1.8677 1.3666
No log 0.3023 26 1.8738 0.2951 1.8738 1.3689
No log 0.3256 28 1.9105 0.2656 1.9105 1.3822
No log 0.3488 30 2.3289 0.1948 2.3289 1.5261
No log 0.3721 32 3.0485 0.0973 3.0485 1.7460
No log 0.3953 34 2.6915 0.1657 2.6915 1.6406
No log 0.4186 36 2.1572 0.2778 2.1572 1.4687
No log 0.4419 38 1.9552 0.2958 1.9552 1.3983
No log 0.4651 40 1.6829 0.3556 1.6829 1.2973
No log 0.4884 42 1.7847 0.3688 1.7847 1.3359
No log 0.5116 44 2.5935 0.2637 2.5935 1.6104
No log 0.5349 46 3.0697 0.2136 3.0697 1.7521
No log 0.5581 48 2.6044 0.2784 2.6044 1.6138
No log 0.5814 50 1.7888 0.4027 1.7888 1.3374
No log 0.6047 52 1.4786 0.4234 1.4786 1.2160
No log 0.6279 54 1.3988 0.4118 1.3988 1.1827
No log 0.6512 56 1.5591 0.4342 1.5591 1.2486
No log 0.6744 58 1.8523 0.4444 1.8523 1.3610
No log 0.6977 60 2.2845 0.3518 2.2845 1.5115
No log 0.7209 62 2.3085 0.3518 2.3085 1.5194
No log 0.7442 64 1.7468 0.4390 1.7468 1.3217
No log 0.7674 66 1.5372 0.4371 1.5372 1.2398
No log 0.7907 68 1.3682 0.4672 1.3682 1.1697
No log 0.8140 70 1.2000 0.4394 1.2000 1.0954
No log 0.8372 72 1.2000 0.4928 1.2000 1.0955
No log 0.8605 74 1.1761 0.4928 1.1761 1.0845
No log 0.8837 76 1.2252 0.5235 1.2252 1.1069
No log 0.9070 78 1.1528 0.5306 1.1528 1.0737
No log 0.9302 80 1.2461 0.5588 1.2461 1.1163
No log 0.9535 82 1.4374 0.4526 1.4374 1.1989
No log 0.9767 84 1.8045 0.3824 1.8045 1.3433
No log 1.0 86 2.0864 0.3043 2.0864 1.4444
No log 1.0233 88 1.6521 0.4276 1.6521 1.2854
No log 1.0465 90 1.4138 0.4762 1.4138 1.1890
No log 1.0698 92 1.3945 0.4698 1.3945 1.1809
No log 1.0930 94 1.4050 0.5059 1.4050 1.1853
No log 1.1163 96 1.2043 0.5503 1.2043 1.0974
No log 1.1395 98 1.0665 0.6447 1.0665 1.0327
No log 1.1628 100 0.9289 0.6853 0.9289 0.9638
No log 1.1860 102 0.8667 0.6714 0.8667 0.9310
No log 1.2093 104 0.9596 0.6131 0.9596 0.9796
No log 1.2326 106 1.0143 0.6043 1.0143 1.0071
No log 1.2558 108 1.0076 0.5811 1.0076 1.0038
No log 1.2791 110 1.1308 0.6437 1.1308 1.0634
No log 1.3023 112 1.1854 0.6437 1.1854 1.0888
No log 1.3256 114 1.0026 0.6581 1.0026 1.0013
No log 1.3488 116 1.2347 0.4962 1.2347 1.1112
No log 1.3721 118 1.2319 0.4580 1.2319 1.1099
No log 1.3953 120 0.9769 0.6809 0.9769 0.9884
No log 1.4186 122 1.0473 0.5594 1.0473 1.0234
No log 1.4419 124 1.5104 0.5422 1.5104 1.2290
No log 1.4651 126 1.6539 0.5652 1.6539 1.2861
No log 1.4884 128 1.7611 0.5178 1.7611 1.3271
No log 1.5116 130 1.3373 0.5889 1.3373 1.1564
No log 1.5349 132 0.9954 0.675 0.9954 0.9977
No log 1.5581 134 0.9561 0.6577 0.9561 0.9778
No log 1.5814 136 0.8548 0.6800 0.8548 0.9245
No log 1.6047 138 0.7988 0.7347 0.7988 0.8937
No log 1.6279 140 0.8063 0.7260 0.8063 0.8980
No log 1.6512 142 0.8574 0.7050 0.8574 0.9260
No log 1.6744 144 0.8667 0.6471 0.8667 0.9310
No log 1.6977 146 0.9518 0.6043 0.9518 0.9756
No log 1.7209 148 1.2604 0.5890 1.2604 1.1227
No log 1.7442 150 1.1710 0.6220 1.1710 1.0821
No log 1.7674 152 0.8806 0.6944 0.8806 0.9384
No log 1.7907 154 0.9354 0.6383 0.9354 0.9672
No log 1.8140 156 0.9341 0.6383 0.9341 0.9665
No log 1.8372 158 0.8217 0.7483 0.8217 0.9065
No log 1.8605 160 0.8235 0.7152 0.8235 0.9075
No log 1.8837 162 0.8404 0.7067 0.8404 0.9168
No log 1.9070 164 1.0135 0.6867 1.0135 1.0067
No log 1.9302 166 0.9277 0.7044 0.9277 0.9632
No log 1.9535 168 0.7775 0.7483 0.7775 0.8818
No log 1.9767 170 0.8253 0.7123 0.8253 0.9085
No log 2.0 172 0.8583 0.7172 0.8583 0.9265
No log 2.0233 174 0.9866 0.6294 0.9866 0.9933
No log 2.0465 176 0.9988 0.6294 0.9988 0.9994
No log 2.0698 178 0.9030 0.6331 0.9030 0.9503
No log 2.0930 180 1.1081 0.5512 1.1081 1.0526
No log 2.1163 182 1.1775 0.5079 1.1775 1.0851
No log 2.1395 184 1.0642 0.5693 1.0642 1.0316
No log 2.1628 186 0.9339 0.6286 0.9339 0.9664
No log 2.1860 188 0.7740 0.76 0.7740 0.8798
No log 2.2093 190 0.7640 0.76 0.7640 0.8741
No log 2.2326 192 0.7728 0.7397 0.7728 0.8791
No log 2.2558 194 0.7235 0.7733 0.7235 0.8506
No log 2.2791 196 0.8083 0.7329 0.8083 0.8990
No log 2.3023 198 1.0623 0.6424 1.0623 1.0307
No log 2.3256 200 0.9732 0.6497 0.9732 0.9865
No log 2.3488 202 0.8250 0.7355 0.8250 0.9083
No log 2.3721 204 0.7252 0.7448 0.7252 0.8516
No log 2.3953 206 0.7307 0.7448 0.7307 0.8548
No log 2.4186 208 0.7369 0.7448 0.7369 0.8584
No log 2.4419 210 0.8438 0.7226 0.8438 0.9186
No log 2.4651 212 0.8246 0.7226 0.8246 0.9081
No log 2.4884 214 0.7947 0.7436 0.7947 0.8914
No log 2.5116 216 0.7182 0.7550 0.7182 0.8475
No log 2.5349 218 0.7190 0.75 0.7190 0.8480
No log 2.5581 220 0.7286 0.7534 0.7286 0.8536
No log 2.5814 222 0.7319 0.7853 0.7319 0.8555
No log 2.6047 224 0.9092 0.7232 0.9092 0.9535
No log 2.6279 226 0.9222 0.6971 0.9222 0.9603
No log 2.6512 228 0.8633 0.7160 0.8633 0.9291
No log 2.6744 230 0.8549 0.6434 0.8549 0.9246
No log 2.6977 232 0.8175 0.6986 0.8175 0.9042
No log 2.7209 234 0.8120 0.6759 0.8120 0.9011
No log 2.7442 236 0.8325 0.6759 0.8325 0.9124
No log 2.7674 238 0.7906 0.6761 0.7906 0.8892
No log 2.7907 240 0.8439 0.6176 0.8439 0.9186
No log 2.8140 242 0.7759 0.7465 0.7759 0.8808
No log 2.8372 244 0.7580 0.7310 0.7580 0.8706
No log 2.8605 246 0.7574 0.7619 0.7574 0.8703
No log 2.8837 248 0.7660 0.7259 0.7660 0.8752
No log 2.9070 250 0.7403 0.7376 0.7403 0.8604
No log 2.9302 252 0.7087 0.75 0.7087 0.8418
No log 2.9535 254 0.7871 0.7950 0.7871 0.8872
No log 2.9767 256 1.0804 0.6215 1.0804 1.0394
No log 3.0 258 1.0441 0.6556 1.0441 1.0218
No log 3.0233 260 0.7890 0.7586 0.7890 0.8883
No log 3.0465 262 0.6987 0.8101 0.6987 0.8359
No log 3.0698 264 0.7408 0.7260 0.7408 0.8607
No log 3.0930 266 0.7429 0.7260 0.7429 0.8619
No log 3.1163 268 0.7966 0.7211 0.7966 0.8925
No log 3.1395 270 0.8970 0.6759 0.8970 0.9471
No log 3.1628 272 0.8621 0.7083 0.8621 0.9285
No log 3.1860 274 0.7669 0.7432 0.7669 0.8758
No log 3.2093 276 0.7525 0.7483 0.7525 0.8675
No log 3.2326 278 0.7208 0.7651 0.7208 0.8490
No log 3.2558 280 0.6996 0.7763 0.6996 0.8364
No log 3.2791 282 0.7091 0.7778 0.7091 0.8421
No log 3.3023 284 0.6790 0.7898 0.6790 0.8240
No log 3.3256 286 0.7430 0.7172 0.7430 0.8620
No log 3.3488 288 0.9431 0.6176 0.9431 0.9711
No log 3.3721 290 0.9969 0.6061 0.9969 0.9984
No log 3.3953 292 0.8710 0.6324 0.8710 0.9333
No log 3.4186 294 0.8159 0.6619 0.8159 0.9032
No log 3.4419 296 0.8156 0.6714 0.8156 0.9031
No log 3.4651 298 0.8070 0.6619 0.8070 0.8984
No log 3.4884 300 0.7838 0.7172 0.7838 0.8853
No log 3.5116 302 0.7649 0.7483 0.7649 0.8746
No log 3.5349 304 0.8219 0.7117 0.8219 0.9066
No log 3.5581 306 0.8618 0.7209 0.8618 0.9283
No log 3.5814 308 0.7978 0.7586 0.7978 0.8932
No log 3.6047 310 0.6648 0.7742 0.6648 0.8154
No log 3.6279 312 0.6343 0.7703 0.6343 0.7964
No log 3.6512 314 0.6481 0.7483 0.6481 0.8050
No log 3.6744 316 0.6741 0.7483 0.6741 0.8210
No log 3.6977 318 0.7242 0.7432 0.7242 0.8510
No log 3.7209 320 0.8138 0.6232 0.8138 0.9021
No log 3.7442 322 0.8209 0.7034 0.8209 0.9061
No log 3.7674 324 0.8519 0.6986 0.8519 0.9230
No log 3.7907 326 0.8428 0.6667 0.8428 0.9181
No log 3.8140 328 0.9061 0.625 0.9061 0.9519
No log 3.8372 330 0.8089 0.7075 0.8089 0.8994
No log 3.8605 332 0.7229 0.7333 0.7229 0.8502
No log 3.8837 334 0.6612 0.7397 0.6612 0.8131
No log 3.9070 336 0.7223 0.7550 0.7223 0.8499
No log 3.9302 338 0.8394 0.6892 0.8394 0.9162
No log 3.9535 340 0.7153 0.7662 0.7153 0.8458
No log 3.9767 342 0.6350 0.7448 0.6350 0.7969
No log 4.0 344 0.6705 0.7376 0.6705 0.8188
No log 4.0233 346 0.7278 0.7050 0.7278 0.8531
No log 4.0465 348 0.7803 0.7101 0.7803 0.8833
No log 4.0698 350 0.7929 0.7042 0.7929 0.8905
No log 4.0930 352 0.7954 0.6950 0.7954 0.8918
No log 4.1163 354 0.8614 0.6712 0.8614 0.9281
No log 4.1395 356 0.8199 0.7114 0.8199 0.9055
No log 4.1628 358 0.7479 0.7310 0.7479 0.8648
No log 4.1860 360 0.7482 0.7361 0.7482 0.8650
No log 4.2093 362 0.7884 0.7297 0.7884 0.8879
No log 4.2326 364 0.8991 0.6803 0.8991 0.9482
No log 4.2558 366 0.9318 0.6712 0.9318 0.9653
No log 4.2791 368 0.8379 0.6897 0.8379 0.9154
No log 4.3023 370 0.8511 0.6331 0.8511 0.9226
No log 4.3256 372 0.9142 0.6232 0.9142 0.9562
No log 4.3488 374 0.8086 0.6950 0.8086 0.8992
No log 4.3721 376 0.7122 0.6809 0.7122 0.8439
No log 4.3953 378 0.7045 0.6950 0.7045 0.8394
No log 4.4186 380 0.7171 0.7183 0.7171 0.8468
No log 4.4419 382 0.7327 0.7092 0.7327 0.8560
No log 4.4651 384 0.7629 0.6906 0.7629 0.8734
No log 4.4884 386 0.7902 0.6906 0.7902 0.8889
No log 4.5116 388 0.8103 0.6906 0.8103 0.9002
No log 4.5349 390 0.7645 0.6857 0.7645 0.8744
No log 4.5581 392 0.7527 0.7682 0.7527 0.8676
No log 4.5814 394 0.8013 0.7317 0.8013 0.8952
No log 4.6047 396 0.7767 0.7613 0.7767 0.8813
No log 4.6279 398 0.7491 0.7632 0.7491 0.8655
No log 4.6512 400 0.7558 0.7467 0.7558 0.8694
No log 4.6744 402 0.7361 0.7682 0.7361 0.8580
No log 4.6977 404 0.7515 0.75 0.7515 0.8669
No log 4.7209 406 0.8150 0.75 0.8150 0.9028
No log 4.7442 408 0.8325 0.7294 0.8325 0.9124
No log 4.7674 410 0.7761 0.7665 0.7761 0.8810
No log 4.7907 412 0.8545 0.6842 0.8545 0.9244
No log 4.8140 414 1.0428 0.6014 1.0428 1.0212
No log 4.8372 416 1.1222 0.5507 1.1222 1.0593
No log 4.8605 418 1.1090 0.5255 1.1090 1.0531
No log 4.8837 420 0.8080 0.7133 0.8080 0.8989
No log 4.9070 422 0.6121 0.75 0.6121 0.7824
No log 4.9302 424 0.5943 0.7867 0.5943 0.7709
No log 4.9535 426 0.6047 0.8 0.6047 0.7776
No log 4.9767 428 0.6452 0.7919 0.6452 0.8033
No log 5.0 430 0.6828 0.7755 0.6828 0.8263
No log 5.0233 432 0.6657 0.7619 0.6657 0.8159
No log 5.0465 434 0.6791 0.7534 0.6791 0.8241
No log 5.0698 436 0.6974 0.7755 0.6974 0.8351
No log 5.0930 438 0.6802 0.8052 0.6802 0.8248
No log 5.1163 440 0.7529 0.7636 0.7529 0.8677
No log 5.1395 442 0.7329 0.7831 0.7329 0.8561
No log 5.1628 444 0.6842 0.8280 0.6842 0.8272
No log 5.1860 446 0.7148 0.7875 0.7148 0.8455
No log 5.2093 448 0.6530 0.8026 0.6530 0.8081
No log 5.2326 450 0.6484 0.7919 0.6484 0.8052
No log 5.2558 452 0.7111 0.7755 0.7111 0.8433
No log 5.2791 454 0.7534 0.7651 0.7534 0.8680
No log 5.3023 456 0.6590 0.7838 0.6590 0.8118
No log 5.3256 458 0.6042 0.75 0.6042 0.7773
No log 5.3488 460 0.6055 0.7413 0.6055 0.7782
No log 5.3721 462 0.6236 0.7619 0.6236 0.7897
No log 5.3953 464 0.6839 0.8 0.6839 0.8270
No log 5.4186 466 0.7462 0.7083 0.7462 0.8638
No log 5.4419 468 0.8124 0.6993 0.8124 0.9014
No log 5.4651 470 0.8938 0.6176 0.8938 0.9454
No log 5.4884 472 0.9300 0.6479 0.9300 0.9644
No log 5.5116 474 0.8998 0.6479 0.8998 0.9486
No log 5.5349 476 0.8624 0.7034 0.8624 0.9286
No log 5.5581 478 0.8390 0.7034 0.8390 0.9160
No log 5.5814 480 0.8662 0.6792 0.8662 0.9307
No log 5.6047 482 0.9149 0.6296 0.9149 0.9565
No log 5.6279 484 0.8757 0.6951 0.8757 0.9358
No log 5.6512 486 0.7774 0.7425 0.7774 0.8817
No log 5.6744 488 0.6936 0.7815 0.6936 0.8328
No log 5.6977 490 0.7119 0.7950 0.7119 0.8437
No log 5.7209 492 0.8170 0.6957 0.8170 0.9039
No log 5.7442 494 0.8747 0.6452 0.8747 0.9352
No log 5.7674 496 0.8677 0.6154 0.8677 0.9315
No log 5.7907 498 0.8029 0.6901 0.8029 0.8961
0.4007 5.8140 500 0.7108 0.7397 0.7108 0.8431
0.4007 5.8372 502 0.7398 0.7172 0.7398 0.8601
0.4007 5.8605 504 0.8247 0.7034 0.8247 0.9081
0.4007 5.8837 506 0.8840 0.6528 0.8840 0.9402
0.4007 5.9070 508 0.8619 0.6528 0.8619 0.9284
0.4007 5.9302 510 0.7839 0.7211 0.7839 0.8854

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run2_AugV5_k18_task1_organization

Finetuned
(4023)
this model