ArabicNewSplits5_FineTuningAraBERT_run3_AugV5_k9_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7342
  • Qwk: 0.5307
  • Mse: 0.7342
  • Rmse: 0.8568

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0345 2 4.0951 -0.0256 4.0951 2.0236
No log 0.0690 4 2.3006 0.0094 2.3006 1.5168
No log 0.1034 6 2.0508 -0.0725 2.0508 1.4320
No log 0.1379 8 1.5048 -0.0826 1.5048 1.2267
No log 0.1724 10 1.0620 0.0306 1.0620 1.0305
No log 0.2069 12 0.7585 0.1955 0.7585 0.8709
No log 0.2414 14 0.6918 0.2832 0.6918 0.8318
No log 0.2759 16 0.7401 0.2819 0.7401 0.8603
No log 0.3103 18 0.8753 0.2910 0.8753 0.9356
No log 0.3448 20 1.1289 0.2161 1.1289 1.0625
No log 0.3793 22 1.1959 0.2132 1.1959 1.0936
No log 0.4138 24 1.5653 0.1991 1.5653 1.2511
No log 0.4483 26 1.8065 0.1308 1.8065 1.3441
No log 0.4828 28 1.3293 0.2114 1.3293 1.1530
No log 0.5172 30 0.8851 0.2663 0.8851 0.9408
No log 0.5517 32 0.8288 0.2531 0.8288 0.9104
No log 0.5862 34 0.8528 0.2387 0.8528 0.9235
No log 0.6207 36 0.8260 0.2531 0.8260 0.9088
No log 0.6552 38 0.6680 0.3491 0.6680 0.8173
No log 0.6897 40 0.5763 0.4 0.5763 0.7591
No log 0.7241 42 0.5920 0.4125 0.5920 0.7694
No log 0.7586 44 0.8732 0.3115 0.8732 0.9344
No log 0.7931 46 1.4194 0.1940 1.4194 1.1914
No log 0.8276 48 1.5596 0.2118 1.5596 1.2488
No log 0.8621 50 1.2691 0.2838 1.2691 1.1265
No log 0.8966 52 0.9218 0.3539 0.9218 0.9601
No log 0.9310 54 0.7146 0.4223 0.7146 0.8453
No log 0.9655 56 0.7739 0.4129 0.7739 0.8797
No log 1.0 58 0.8080 0.4085 0.8080 0.8989
No log 1.0345 60 0.9883 0.3528 0.9883 0.9941
No log 1.0690 62 1.1114 0.3339 1.1114 1.0542
No log 1.1034 64 1.0591 0.3682 1.0591 1.0291
No log 1.1379 66 0.8499 0.4282 0.8499 0.9219
No log 1.1724 68 0.7122 0.4077 0.7122 0.8439
No log 1.2069 70 0.6947 0.3965 0.6947 0.8335
No log 1.2414 72 0.7382 0.3992 0.7382 0.8592
No log 1.2759 74 1.1177 0.3996 1.1177 1.0572
No log 1.3103 76 1.5130 0.2707 1.5130 1.2300
No log 1.3448 78 1.3798 0.3639 1.3798 1.1746
No log 1.3793 80 1.0569 0.4617 1.0569 1.0281
No log 1.4138 82 0.8527 0.4312 0.8527 0.9234
No log 1.4483 84 0.7077 0.4805 0.7077 0.8412
No log 1.4828 86 0.7217 0.4713 0.7217 0.8496
No log 1.5172 88 0.8254 0.4695 0.8254 0.9085
No log 1.5517 90 1.1039 0.4272 1.1039 1.0507
No log 1.5862 92 1.1137 0.4037 1.1137 1.0553
No log 1.6207 94 0.8690 0.5121 0.8690 0.9322
No log 1.6552 96 0.7343 0.5408 0.7343 0.8569
No log 1.6897 98 0.7069 0.5324 0.7069 0.8408
No log 1.7241 100 0.6999 0.4931 0.6999 0.8366
No log 1.7586 102 0.8220 0.4460 0.8220 0.9067
No log 1.7931 104 0.7515 0.4735 0.7515 0.8669
No log 1.8276 106 0.6721 0.471 0.6721 0.8198
No log 1.8621 108 0.7036 0.4847 0.7036 0.8388
No log 1.8966 110 0.7991 0.4203 0.7991 0.8939
No log 1.9310 112 0.9099 0.4194 0.9099 0.9539
No log 1.9655 114 0.8442 0.4438 0.8442 0.9188
No log 2.0 116 0.7452 0.5285 0.7452 0.8632
No log 2.0345 118 0.7381 0.5101 0.7381 0.8592
No log 2.0690 120 0.7432 0.5436 0.7432 0.8621
No log 2.1034 122 0.8826 0.5012 0.8826 0.9395
No log 2.1379 124 1.1477 0.3928 1.1477 1.0713
No log 2.1724 126 1.0992 0.3993 1.0992 1.0484
No log 2.2069 128 0.8532 0.4785 0.8532 0.9237
No log 2.2414 130 0.6833 0.4695 0.6833 0.8266
No log 2.2759 132 0.6528 0.4736 0.6528 0.8080
No log 2.3103 134 0.6665 0.4991 0.6665 0.8164
No log 2.3448 136 0.6806 0.4551 0.6806 0.8250
No log 2.3793 138 0.7455 0.4533 0.7455 0.8634
No log 2.4138 140 0.8139 0.4770 0.8139 0.9022
No log 2.4483 142 0.8823 0.4829 0.8823 0.9393
No log 2.4828 144 0.9378 0.4922 0.9378 0.9684
No log 2.5172 146 0.8842 0.5055 0.8842 0.9403
No log 2.5517 148 0.8960 0.4722 0.8960 0.9466
No log 2.5862 150 0.9844 0.4891 0.9844 0.9922
No log 2.6207 152 0.9841 0.4479 0.9841 0.9920
No log 2.6552 154 0.7954 0.4840 0.7954 0.8918
No log 2.6897 156 0.7023 0.5037 0.7023 0.8380
No log 2.7241 158 0.7042 0.4779 0.7042 0.8392
No log 2.7586 160 0.7384 0.4627 0.7384 0.8593
No log 2.7931 162 0.8421 0.4882 0.8421 0.9177
No log 2.8276 164 1.0093 0.4379 1.0093 1.0047
No log 2.8621 166 1.0016 0.4247 1.0016 1.0008
No log 2.8966 168 0.8771 0.4841 0.8771 0.9365
No log 2.9310 170 0.8059 0.4408 0.8059 0.8977
No log 2.9655 172 0.8554 0.4871 0.8554 0.9249
No log 3.0 174 1.0395 0.4051 1.0395 1.0195
No log 3.0345 176 1.2574 0.3484 1.2574 1.1213
No log 3.0690 178 1.1934 0.3560 1.1934 1.0924
No log 3.1034 180 0.9877 0.4149 0.9877 0.9938
No log 3.1379 182 0.8105 0.4906 0.8105 0.9003
No log 3.1724 184 0.7298 0.4537 0.7298 0.8543
No log 3.2069 186 0.7254 0.4684 0.7254 0.8517
No log 3.2414 188 0.7835 0.4468 0.7835 0.8852
No log 3.2759 190 0.8313 0.4609 0.8313 0.9117
No log 3.3103 192 0.9163 0.3926 0.9163 0.9572
No log 3.3448 194 0.8475 0.4394 0.8475 0.9206
No log 3.3793 196 0.7743 0.4575 0.7743 0.8800
No log 3.4138 198 0.7515 0.4873 0.7515 0.8669
No log 3.4483 200 0.8119 0.4062 0.8119 0.9011
No log 3.4828 202 0.8565 0.4438 0.8565 0.9255
No log 3.5172 204 0.8981 0.4639 0.8981 0.9477
No log 3.5517 206 0.8197 0.4290 0.8197 0.9054
No log 3.5862 208 0.7915 0.4441 0.7915 0.8897
No log 3.6207 210 0.7477 0.4690 0.7477 0.8647
No log 3.6552 212 0.8145 0.4121 0.8145 0.9025
No log 3.6897 214 0.8998 0.4440 0.8998 0.9486
No log 3.7241 216 0.8845 0.4437 0.8845 0.9405
No log 3.7586 218 0.8055 0.4932 0.8055 0.8975
No log 3.7931 220 0.7321 0.5235 0.7321 0.8556
No log 3.8276 222 0.7314 0.5175 0.7314 0.8552
No log 3.8621 224 0.7476 0.5179 0.7476 0.8646
No log 3.8966 226 0.7870 0.4952 0.7870 0.8871
No log 3.9310 228 0.8445 0.5283 0.8445 0.9189
No log 3.9655 230 0.8026 0.5033 0.8026 0.8959
No log 4.0 232 0.7561 0.5538 0.7561 0.8695
No log 4.0345 234 0.7612 0.4720 0.7612 0.8725
No log 4.0690 236 0.7494 0.4825 0.7494 0.8657
No log 4.1034 238 0.7593 0.4770 0.7593 0.8714
No log 4.1379 240 0.7847 0.5039 0.7847 0.8858
No log 4.1724 242 0.8392 0.4956 0.8392 0.9161
No log 4.2069 244 0.8524 0.4964 0.8524 0.9233
No log 4.2414 246 0.7969 0.5030 0.7969 0.8927
No log 4.2759 248 0.7738 0.5238 0.7738 0.8797
No log 4.3103 250 0.7268 0.5035 0.7268 0.8525
No log 4.3448 252 0.6763 0.5474 0.6763 0.8224
No log 4.3793 254 0.6659 0.4845 0.6659 0.8160
No log 4.4138 256 0.6478 0.5309 0.6478 0.8048
No log 4.4483 258 0.6503 0.5276 0.6503 0.8064
No log 4.4828 260 0.7001 0.4792 0.7001 0.8367
No log 4.5172 262 0.8172 0.3955 0.8172 0.9040
No log 4.5517 264 0.8514 0.3842 0.8514 0.9227
No log 4.5862 266 0.7889 0.4294 0.7889 0.8882
No log 4.6207 268 0.7102 0.5174 0.7102 0.8427
No log 4.6552 270 0.6953 0.5835 0.6953 0.8338
No log 4.6897 272 0.7128 0.5463 0.7128 0.8443
No log 4.7241 274 0.7227 0.5239 0.7227 0.8501
No log 4.7586 276 0.7161 0.5332 0.7161 0.8462
No log 4.7931 278 0.7221 0.5230 0.7221 0.8498
No log 4.8276 280 0.7244 0.5326 0.7244 0.8511
No log 4.8621 282 0.7501 0.5455 0.7501 0.8661
No log 4.8966 284 0.7341 0.5468 0.7341 0.8568
No log 4.9310 286 0.6886 0.5331 0.6886 0.8298
No log 4.9655 288 0.6593 0.5363 0.6593 0.8120
No log 5.0 290 0.6527 0.5060 0.6527 0.8079
No log 5.0345 292 0.6433 0.4964 0.6433 0.8020
No log 5.0690 294 0.6218 0.5347 0.6218 0.7885
No log 5.1034 296 0.6503 0.5585 0.6503 0.8064
No log 5.1379 298 0.6935 0.5422 0.6935 0.8328
No log 5.1724 300 0.6733 0.5405 0.6733 0.8206
No log 5.2069 302 0.6355 0.5792 0.6355 0.7972
No log 5.2414 304 0.6273 0.5773 0.6273 0.7920
No log 5.2759 306 0.6236 0.5516 0.6236 0.7897
No log 5.3103 308 0.6296 0.5836 0.6296 0.7935
No log 5.3448 310 0.6558 0.5709 0.6558 0.8098
No log 5.3793 312 0.7087 0.5746 0.7087 0.8419
No log 5.4138 314 0.7773 0.5684 0.7773 0.8817
No log 5.4483 316 0.7777 0.5584 0.7777 0.8819
No log 5.4828 318 0.7472 0.5665 0.7472 0.8644
No log 5.5172 320 0.7399 0.5038 0.7399 0.8602
No log 5.5517 322 0.7529 0.5127 0.7529 0.8677
No log 5.5862 324 0.7270 0.5465 0.7270 0.8526
No log 5.6207 326 0.7045 0.5738 0.7045 0.8394
No log 5.6552 328 0.7180 0.5689 0.7180 0.8473
No log 5.6897 330 0.7307 0.5572 0.7307 0.8548
No log 5.7241 332 0.7391 0.5572 0.7391 0.8597
No log 5.7586 334 0.6948 0.5616 0.6948 0.8335
No log 5.7931 336 0.6841 0.5781 0.6841 0.8271
No log 5.8276 338 0.6866 0.5812 0.6866 0.8286
No log 5.8621 340 0.6850 0.5718 0.6850 0.8276
No log 5.8966 342 0.6838 0.5718 0.6838 0.8269
No log 5.9310 344 0.6923 0.5733 0.6923 0.8321
No log 5.9655 346 0.7100 0.5539 0.7100 0.8426
No log 6.0 348 0.7333 0.5844 0.7333 0.8563
No log 6.0345 350 0.7436 0.5311 0.7436 0.8623
No log 6.0690 352 0.7305 0.5556 0.7305 0.8547
No log 6.1034 354 0.7412 0.5557 0.7412 0.8609
No log 6.1379 356 0.7430 0.5382 0.7430 0.8620
No log 6.1724 358 0.7773 0.4942 0.7773 0.8816
No log 6.2069 360 0.7735 0.4827 0.7735 0.8795
No log 6.2414 362 0.7794 0.4756 0.7794 0.8828
No log 6.2759 364 0.7639 0.5111 0.7639 0.8740
No log 6.3103 366 0.7126 0.5721 0.7126 0.8442
No log 6.3448 368 0.6940 0.5788 0.6940 0.8331
No log 6.3793 370 0.6904 0.5788 0.6904 0.8309
No log 6.4138 372 0.7051 0.5692 0.7051 0.8397
No log 6.4483 374 0.7096 0.5692 0.7096 0.8424
No log 6.4828 376 0.7286 0.5658 0.7286 0.8536
No log 6.5172 378 0.7376 0.5673 0.7376 0.8588
No log 6.5517 380 0.7325 0.5328 0.7325 0.8559
No log 6.5862 382 0.7277 0.5104 0.7277 0.8531
No log 6.6207 384 0.7138 0.5104 0.7138 0.8449
No log 6.6552 386 0.7020 0.5551 0.7020 0.8379
No log 6.6897 388 0.6996 0.5878 0.6996 0.8364
No log 6.7241 390 0.7180 0.5346 0.7180 0.8474
No log 6.7586 392 0.7362 0.5121 0.7362 0.8580
No log 6.7931 394 0.7196 0.5121 0.7196 0.8483
No log 6.8276 396 0.6688 0.5556 0.6688 0.8178
No log 6.8621 398 0.6332 0.5804 0.6332 0.7957
No log 6.8966 400 0.6423 0.5340 0.6423 0.8014
No log 6.9310 402 0.6778 0.4846 0.6778 0.8233
No log 6.9655 404 0.6841 0.4788 0.6841 0.8271
No log 7.0 406 0.6778 0.4973 0.6778 0.8233
No log 7.0345 408 0.6951 0.5831 0.6951 0.8338
No log 7.0690 410 0.7506 0.5350 0.7506 0.8664
No log 7.1034 412 0.7725 0.5300 0.7725 0.8789
No log 7.1379 414 0.7661 0.5599 0.7661 0.8753
No log 7.1724 416 0.7426 0.5907 0.7426 0.8618
No log 7.2069 418 0.7258 0.5827 0.7258 0.8519
No log 7.2414 420 0.7041 0.5269 0.7041 0.8391
No log 7.2759 422 0.6890 0.4934 0.6890 0.8301
No log 7.3103 424 0.6758 0.5628 0.6758 0.8221
No log 7.3448 426 0.6745 0.5898 0.6745 0.8213
No log 7.3793 428 0.6738 0.5831 0.6738 0.8208
No log 7.4138 430 0.6703 0.5835 0.6703 0.8187
No log 7.4483 432 0.6670 0.5898 0.6670 0.8167
No log 7.4828 434 0.6657 0.5888 0.6657 0.8159
No log 7.5172 436 0.6809 0.5097 0.6809 0.8251
No log 7.5517 438 0.7091 0.4882 0.7091 0.8421
No log 7.5862 440 0.7292 0.4589 0.7292 0.8539
No log 7.6207 442 0.7250 0.4589 0.7250 0.8515
No log 7.6552 444 0.7057 0.5068 0.7057 0.8401
No log 7.6897 446 0.6885 0.5672 0.6885 0.8298
No log 7.7241 448 0.6916 0.5835 0.6916 0.8316
No log 7.7586 450 0.6963 0.5524 0.6963 0.8344
No log 7.7931 452 0.6982 0.5645 0.6982 0.8356
No log 7.8276 454 0.6924 0.5964 0.6924 0.8321
No log 7.8621 456 0.6917 0.5489 0.6917 0.8317
No log 7.8966 458 0.6916 0.5236 0.6916 0.8316
No log 7.9310 460 0.6846 0.5430 0.6846 0.8274
No log 7.9655 462 0.6784 0.5364 0.6784 0.8236
No log 8.0 464 0.6777 0.5534 0.6777 0.8232
No log 8.0345 466 0.6807 0.5611 0.6807 0.8251
No log 8.0690 468 0.6742 0.5611 0.6742 0.8211
No log 8.1034 470 0.6728 0.5566 0.6728 0.8202
No log 8.1379 472 0.6735 0.5582 0.6735 0.8207
No log 8.1724 474 0.6781 0.5893 0.6781 0.8235
No log 8.2069 476 0.6844 0.5827 0.6844 0.8273
No log 8.2414 478 0.6913 0.5569 0.6913 0.8314
No log 8.2759 480 0.6915 0.5766 0.6915 0.8316
No log 8.3103 482 0.6877 0.5827 0.6877 0.8293
No log 8.3448 484 0.6875 0.5521 0.6875 0.8292
No log 8.3793 486 0.6841 0.5628 0.6841 0.8271
No log 8.4138 488 0.6851 0.5358 0.6851 0.8277
No log 8.4483 490 0.6894 0.5032 0.6894 0.8303
No log 8.4828 492 0.6826 0.5033 0.6826 0.8262
No log 8.5172 494 0.6823 0.5092 0.6823 0.8260
No log 8.5517 496 0.6748 0.5033 0.6748 0.8215
No log 8.5862 498 0.6726 0.5459 0.6726 0.8202
0.3803 8.6207 500 0.6732 0.5295 0.6732 0.8205
0.3803 8.6552 502 0.6740 0.5644 0.6740 0.8210
0.3803 8.6897 504 0.6781 0.5582 0.6781 0.8235
0.3803 8.7241 506 0.6821 0.5733 0.6821 0.8259
0.3803 8.7586 508 0.6807 0.5736 0.6807 0.8250
0.3803 8.7931 510 0.6805 0.5733 0.6805 0.8249
0.3803 8.8276 512 0.6863 0.5629 0.6863 0.8285
0.3803 8.8621 514 0.6924 0.5644 0.6924 0.8321
0.3803 8.8966 516 0.6951 0.5644 0.6951 0.8337
0.3803 8.9310 518 0.6902 0.5629 0.6902 0.8308
0.3803 8.9655 520 0.6868 0.5690 0.6868 0.8288
0.3803 9.0 522 0.6810 0.5538 0.6810 0.8252
0.3803 9.0345 524 0.6790 0.5567 0.6790 0.8240
0.3803 9.0690 526 0.6807 0.5567 0.6807 0.8250
0.3803 9.1034 528 0.6815 0.5718 0.6815 0.8255
0.3803 9.1379 530 0.6850 0.5567 0.6850 0.8276
0.3803 9.1724 532 0.6850 0.5566 0.6850 0.8277
0.3803 9.2069 534 0.6844 0.5551 0.6844 0.8273
0.3803 9.2414 536 0.6841 0.5551 0.6841 0.8271
0.3803 9.2759 538 0.6858 0.5551 0.6858 0.8282
0.3803 9.3103 540 0.6872 0.5551 0.6872 0.8290
0.3803 9.3448 542 0.6909 0.5360 0.6909 0.8312
0.3803 9.3793 544 0.6941 0.5360 0.6941 0.8331
0.3803 9.4138 546 0.6970 0.5360 0.6970 0.8349
0.3803 9.4483 548 0.6995 0.5654 0.6995 0.8364
0.3803 9.4828 550 0.7020 0.5579 0.7020 0.8378
0.3803 9.5172 552 0.7053 0.5358 0.7053 0.8398
0.3803 9.5517 554 0.7100 0.5416 0.7100 0.8426
0.3803 9.5862 556 0.7158 0.5403 0.7158 0.8460
0.3803 9.6207 558 0.7216 0.5403 0.7216 0.8495
0.3803 9.6552 560 0.7254 0.5390 0.7254 0.8517
0.3803 9.6897 562 0.7287 0.5390 0.7287 0.8536
0.3803 9.7241 564 0.7313 0.5390 0.7313 0.8551
0.3803 9.7586 566 0.7337 0.5403 0.7337 0.8565
0.3803 9.7931 568 0.7342 0.5348 0.7342 0.8568
0.3803 9.8276 570 0.7352 0.5348 0.7352 0.8574
0.3803 9.8621 572 0.7352 0.5292 0.7352 0.8575
0.3803 9.8966 574 0.7349 0.5296 0.7349 0.8573
0.3803 9.9310 576 0.7346 0.5296 0.7346 0.8571
0.3803 9.9655 578 0.7343 0.5307 0.7343 0.8569
0.3803 10.0 580 0.7342 0.5307 0.7342 0.8568

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run3_AugV5_k9_task2_organization

Finetuned
(4023)
this model