ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k9_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7358
  • Qwk: 0.5296
  • Mse: 0.7358
  • Rmse: 0.8578

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0345 2 4.0951 -0.0256 4.0951 2.0236
No log 0.0690 4 2.3006 0.0094 2.3006 1.5168
No log 0.1034 6 2.0508 -0.0725 2.0508 1.4320
No log 0.1379 8 1.5048 -0.0826 1.5048 1.2267
No log 0.1724 10 1.0620 0.0306 1.0620 1.0305
No log 0.2069 12 0.7585 0.1955 0.7585 0.8709
No log 0.2414 14 0.6918 0.2832 0.6918 0.8318
No log 0.2759 16 0.7401 0.2819 0.7401 0.8603
No log 0.3103 18 0.8753 0.2910 0.8753 0.9356
No log 0.3448 20 1.1289 0.2161 1.1289 1.0625
No log 0.3793 22 1.1959 0.2132 1.1959 1.0936
No log 0.4138 24 1.5653 0.1991 1.5653 1.2511
No log 0.4483 26 1.8065 0.1308 1.8065 1.3441
No log 0.4828 28 1.3293 0.2114 1.3293 1.1530
No log 0.5172 30 0.8851 0.2663 0.8851 0.9408
No log 0.5517 32 0.8288 0.2531 0.8288 0.9104
No log 0.5862 34 0.8528 0.2387 0.8528 0.9235
No log 0.6207 36 0.8260 0.2531 0.8260 0.9088
No log 0.6552 38 0.6680 0.3491 0.6680 0.8173
No log 0.6897 40 0.5763 0.4 0.5763 0.7591
No log 0.7241 42 0.5920 0.4125 0.5920 0.7694
No log 0.7586 44 0.8732 0.3115 0.8732 0.9344
No log 0.7931 46 1.4194 0.1940 1.4194 1.1914
No log 0.8276 48 1.5596 0.2118 1.5596 1.2488
No log 0.8621 50 1.2691 0.2838 1.2691 1.1265
No log 0.8966 52 0.9218 0.3539 0.9218 0.9601
No log 0.9310 54 0.7146 0.4223 0.7146 0.8453
No log 0.9655 56 0.7739 0.4129 0.7739 0.8797
No log 1.0 58 0.8080 0.4085 0.8080 0.8989
No log 1.0345 60 0.9883 0.3528 0.9883 0.9941
No log 1.0690 62 1.1114 0.3339 1.1114 1.0542
No log 1.1034 64 1.0591 0.3682 1.0591 1.0291
No log 1.1379 66 0.8499 0.4282 0.8499 0.9219
No log 1.1724 68 0.7122 0.4077 0.7122 0.8439
No log 1.2069 70 0.6947 0.3965 0.6947 0.8335
No log 1.2414 72 0.7382 0.3992 0.7382 0.8592
No log 1.2759 74 1.1177 0.3996 1.1177 1.0572
No log 1.3103 76 1.5130 0.2707 1.5130 1.2300
No log 1.3448 78 1.3798 0.3639 1.3798 1.1746
No log 1.3793 80 1.0569 0.4617 1.0569 1.0281
No log 1.4138 82 0.8527 0.4312 0.8527 0.9234
No log 1.4483 84 0.7077 0.4805 0.7077 0.8412
No log 1.4828 86 0.7217 0.4713 0.7217 0.8496
No log 1.5172 88 0.8254 0.4695 0.8254 0.9085
No log 1.5517 90 1.1039 0.4272 1.1039 1.0507
No log 1.5862 92 1.1137 0.4037 1.1137 1.0553
No log 1.6207 94 0.8690 0.5121 0.8690 0.9322
No log 1.6552 96 0.7343 0.5408 0.7343 0.8569
No log 1.6897 98 0.7069 0.5324 0.7069 0.8408
No log 1.7241 100 0.6999 0.4931 0.6999 0.8366
No log 1.7586 102 0.8220 0.4460 0.8220 0.9067
No log 1.7931 104 0.7515 0.4735 0.7515 0.8669
No log 1.8276 106 0.6721 0.471 0.6721 0.8198
No log 1.8621 108 0.7036 0.4847 0.7036 0.8388
No log 1.8966 110 0.7991 0.4203 0.7991 0.8939
No log 1.9310 112 0.9099 0.4194 0.9099 0.9539
No log 1.9655 114 0.8442 0.4438 0.8442 0.9188
No log 2.0 116 0.7451 0.5285 0.7451 0.8632
No log 2.0345 118 0.7381 0.5101 0.7381 0.8592
No log 2.0690 120 0.7432 0.5436 0.7432 0.8621
No log 2.1034 122 0.8826 0.5012 0.8826 0.9395
No log 2.1379 124 1.1477 0.3928 1.1477 1.0713
No log 2.1724 126 1.0992 0.3993 1.0992 1.0484
No log 2.2069 128 0.8532 0.4785 0.8532 0.9237
No log 2.2414 130 0.6833 0.4695 0.6833 0.8266
No log 2.2759 132 0.6528 0.4736 0.6528 0.8080
No log 2.3103 134 0.6665 0.4991 0.6665 0.8164
No log 2.3448 136 0.6806 0.4551 0.6806 0.8250
No log 2.3793 138 0.7455 0.4533 0.7455 0.8634
No log 2.4138 140 0.8138 0.4770 0.8138 0.9021
No log 2.4483 142 0.8822 0.4829 0.8822 0.9392
No log 2.4828 144 0.9377 0.4922 0.9377 0.9684
No log 2.5172 146 0.8841 0.5055 0.8841 0.9403
No log 2.5517 148 0.8959 0.4722 0.8959 0.9465
No log 2.5862 150 0.9843 0.4891 0.9843 0.9921
No log 2.6207 152 0.9840 0.4479 0.9840 0.9920
No log 2.6552 154 0.7953 0.4840 0.7953 0.8918
No log 2.6897 156 0.7023 0.5037 0.7023 0.8380
No log 2.7241 158 0.7042 0.4973 0.7042 0.8392
No log 2.7586 160 0.7385 0.4627 0.7385 0.8593
No log 2.7931 162 0.8422 0.4882 0.8422 0.9177
No log 2.8276 164 1.0092 0.4379 1.0092 1.0046
No log 2.8621 166 1.0013 0.4297 1.0013 1.0007
No log 2.8966 168 0.8769 0.4841 0.8769 0.9364
No log 2.9310 170 0.8059 0.4408 0.8059 0.8977
No log 2.9655 172 0.8555 0.4871 0.8555 0.9249
No log 3.0 174 1.0397 0.4051 1.0397 1.0196
No log 3.0345 176 1.2572 0.3484 1.2572 1.1212
No log 3.0690 178 1.1929 0.3560 1.1929 1.0922
No log 3.1034 180 0.9871 0.4149 0.9871 0.9935
No log 3.1379 182 0.8103 0.4906 0.8103 0.9002
No log 3.1724 184 0.7298 0.4537 0.7298 0.8543
No log 3.2069 186 0.7255 0.4684 0.7255 0.8518
No log 3.2414 188 0.7842 0.4547 0.7842 0.8855
No log 3.2759 190 0.8318 0.4609 0.8318 0.9120
No log 3.3103 192 0.9162 0.3926 0.9162 0.9572
No log 3.3448 194 0.8471 0.4394 0.8471 0.9204
No log 3.3793 196 0.7740 0.4575 0.7740 0.8797
No log 3.4138 198 0.7513 0.4873 0.7513 0.8668
No log 3.4483 200 0.8117 0.4062 0.8117 0.9009
No log 3.4828 202 0.8566 0.4438 0.8566 0.9255
No log 3.5172 204 0.8987 0.4639 0.8987 0.9480
No log 3.5517 206 0.8206 0.4290 0.8206 0.9058
No log 3.5862 208 0.7923 0.4458 0.7923 0.8901
No log 3.6207 210 0.7482 0.4690 0.7482 0.8650
No log 3.6552 212 0.8147 0.4121 0.8147 0.9026
No log 3.6897 214 0.8994 0.4440 0.8994 0.9483
No log 3.7241 216 0.8837 0.4437 0.8837 0.9401
No log 3.7586 218 0.8053 0.4932 0.8053 0.8974
No log 3.7931 220 0.7324 0.5235 0.7324 0.8558
No log 3.8276 222 0.7316 0.5175 0.7316 0.8553
No log 3.8621 224 0.7479 0.5179 0.7479 0.8648
No log 3.8966 226 0.7872 0.4952 0.7872 0.8872
No log 3.9310 228 0.8445 0.5283 0.8445 0.9190
No log 3.9655 230 0.8025 0.5033 0.8025 0.8958
No log 4.0 232 0.7558 0.5538 0.7558 0.8694
No log 4.0345 234 0.7609 0.4720 0.7609 0.8723
No log 4.0690 236 0.7492 0.4825 0.7492 0.8656
No log 4.1034 238 0.7592 0.4770 0.7592 0.8713
No log 4.1379 240 0.7850 0.5039 0.7850 0.8860
No log 4.1724 242 0.8396 0.4956 0.8396 0.9163
No log 4.2069 244 0.8526 0.4964 0.8526 0.9234
No log 4.2414 246 0.7970 0.5030 0.7970 0.8927
No log 4.2759 248 0.7739 0.5238 0.7739 0.8797
No log 4.3103 250 0.7267 0.5035 0.7267 0.8524
No log 4.3448 252 0.6762 0.5474 0.6762 0.8223
No log 4.3793 254 0.6659 0.4845 0.6659 0.8161
No log 4.4138 256 0.6476 0.5309 0.6476 0.8047
No log 4.4483 258 0.6499 0.5276 0.6499 0.8061
No log 4.4828 260 0.6995 0.4792 0.6995 0.8363
No log 4.5172 262 0.8174 0.3955 0.8174 0.9041
No log 4.5517 264 0.8527 0.3842 0.8527 0.9234
No log 4.5862 266 0.7900 0.4294 0.7900 0.8888
No log 4.6207 268 0.7101 0.5167 0.7101 0.8427
No log 4.6552 270 0.6947 0.5835 0.6947 0.8335
No log 4.6897 272 0.7123 0.5463 0.7123 0.8440
No log 4.7241 274 0.7223 0.5432 0.7223 0.8499
No log 4.7586 276 0.7161 0.5258 0.7161 0.8463
No log 4.7931 278 0.7222 0.5230 0.7222 0.8498
No log 4.8276 280 0.7250 0.5326 0.7250 0.8514
No log 4.8621 282 0.7505 0.5455 0.7505 0.8663
No log 4.8966 284 0.7341 0.5468 0.7341 0.8568
No log 4.9310 286 0.6884 0.5331 0.6884 0.8297
No log 4.9655 288 0.6590 0.5363 0.6590 0.8118
No log 5.0 290 0.6524 0.5060 0.6524 0.8077
No log 5.0345 292 0.6430 0.4964 0.6430 0.8019
No log 5.0690 294 0.6213 0.5347 0.6213 0.7882
No log 5.1034 296 0.6493 0.5585 0.6493 0.8058
No log 5.1379 298 0.6926 0.5422 0.6926 0.8322
No log 5.1724 300 0.6726 0.5540 0.6726 0.8201
No log 5.2069 302 0.6348 0.5792 0.6348 0.7967
No log 5.2414 304 0.6266 0.5773 0.6266 0.7916
No log 5.2759 306 0.6230 0.5516 0.6230 0.7893
No log 5.3103 308 0.6292 0.5836 0.6292 0.7933
No log 5.3448 310 0.6556 0.5709 0.6556 0.8097
No log 5.3793 312 0.7086 0.5597 0.7086 0.8418
No log 5.4138 314 0.7772 0.5684 0.7772 0.8816
No log 5.4483 316 0.7777 0.5584 0.7777 0.8819
No log 5.4828 318 0.7473 0.5665 0.7473 0.8645
No log 5.5172 320 0.7402 0.5038 0.7402 0.8604
No log 5.5517 322 0.7536 0.5127 0.7536 0.8681
No log 5.5862 324 0.7274 0.5206 0.7274 0.8529
No log 5.6207 326 0.7045 0.5738 0.7045 0.8393
No log 5.6552 328 0.7177 0.5689 0.7177 0.8472
No log 5.6897 330 0.7305 0.5572 0.7305 0.8547
No log 5.7241 332 0.7392 0.5572 0.7392 0.8597
No log 5.7586 334 0.6951 0.5616 0.6951 0.8337
No log 5.7931 336 0.6843 0.5781 0.6843 0.8272
No log 5.8276 338 0.6867 0.5873 0.6867 0.8287
No log 5.8621 340 0.6848 0.5718 0.6848 0.8275
No log 5.8966 342 0.6835 0.5718 0.6835 0.8267
No log 5.9310 344 0.6919 0.5733 0.6919 0.8318
No log 5.9655 346 0.7095 0.5539 0.7095 0.8423
No log 6.0 348 0.7342 0.5844 0.7342 0.8569
No log 6.0345 350 0.7465 0.5311 0.7465 0.8640
No log 6.0690 352 0.7336 0.5556 0.7336 0.8565
No log 6.1034 354 0.7444 0.5557 0.7444 0.8628
No log 6.1379 356 0.7463 0.5382 0.7463 0.8639
No log 6.1724 358 0.7812 0.4827 0.7812 0.8839
No log 6.2069 360 0.7773 0.4827 0.7773 0.8817
No log 6.2414 362 0.7820 0.4756 0.7820 0.8843
No log 6.2759 364 0.7649 0.4989 0.7649 0.8746
No log 6.3103 366 0.7126 0.5601 0.7126 0.8441
No log 6.3448 368 0.6937 0.5788 0.6937 0.8329
No log 6.3793 370 0.6906 0.5788 0.6906 0.8310
No log 6.4138 372 0.7068 0.5709 0.7068 0.8407
No log 6.4483 374 0.7122 0.5692 0.7122 0.8439
No log 6.4828 376 0.7301 0.5658 0.7301 0.8544
No log 6.5172 378 0.7370 0.5537 0.7370 0.8585
No log 6.5517 380 0.7315 0.5316 0.7315 0.8553
No log 6.5862 382 0.7284 0.5104 0.7284 0.8535
No log 6.6207 384 0.7146 0.5104 0.7146 0.8453
No log 6.6552 386 0.7020 0.54 0.7020 0.8378
No log 6.6897 388 0.7003 0.5878 0.7003 0.8368
No log 6.7241 390 0.7215 0.5346 0.7215 0.8494
No log 6.7586 392 0.7436 0.5003 0.7436 0.8623
No log 6.7931 394 0.7293 0.4926 0.7293 0.8540
No log 6.8276 396 0.6766 0.5556 0.6766 0.8226
No log 6.8621 398 0.6361 0.5903 0.6361 0.7975
No log 6.8966 400 0.6409 0.5340 0.6409 0.8006
No log 6.9310 402 0.6777 0.4846 0.6777 0.8232
No log 6.9655 404 0.6862 0.4788 0.6862 0.8283
No log 7.0 406 0.6791 0.4826 0.6791 0.8241
No log 7.0345 408 0.6918 0.5566 0.6918 0.8317
No log 7.0690 410 0.7427 0.5346 0.7427 0.8618
No log 7.1034 412 0.7660 0.5485 0.7660 0.8752
No log 7.1379 414 0.7629 0.5470 0.7629 0.8734
No log 7.1724 416 0.7421 0.5793 0.7421 0.8615
No log 7.2069 418 0.7263 0.5838 0.7263 0.8522
No log 7.2414 420 0.7032 0.5551 0.7032 0.8386
No log 7.2759 422 0.6878 0.5150 0.6878 0.8293
No log 7.3103 424 0.6761 0.5644 0.6761 0.8222
No log 7.3448 426 0.6754 0.5898 0.6754 0.8218
No log 7.3793 428 0.6743 0.5835 0.6743 0.8212
No log 7.4138 430 0.6707 0.5835 0.6707 0.8190
No log 7.4483 432 0.6680 0.5898 0.6680 0.8173
No log 7.4828 434 0.6670 0.5727 0.6670 0.8167
No log 7.5172 436 0.6815 0.5097 0.6815 0.8255
No log 7.5517 438 0.7079 0.4882 0.7079 0.8414
No log 7.5862 440 0.7267 0.4589 0.7267 0.8525
No log 7.6207 442 0.7226 0.4725 0.7226 0.8501
No log 7.6552 444 0.7045 0.5214 0.7045 0.8394
No log 7.6897 446 0.6882 0.5672 0.6882 0.8296
No log 7.7241 448 0.6910 0.5970 0.6910 0.8313
No log 7.7586 450 0.6948 0.5585 0.6948 0.8335
No log 7.7931 452 0.6963 0.5705 0.6963 0.8344
No log 7.8276 454 0.6909 0.5964 0.6909 0.8312
No log 7.8621 456 0.6913 0.5489 0.6913 0.8315
No log 7.8966 458 0.6916 0.5430 0.6916 0.8316
No log 7.9310 460 0.6842 0.5430 0.6842 0.8272
No log 7.9655 462 0.6776 0.5736 0.6776 0.8232
No log 8.0 464 0.6772 0.5534 0.6772 0.8229
No log 8.0345 466 0.6806 0.5611 0.6806 0.8250
No log 8.0690 468 0.6745 0.5566 0.6745 0.8213
No log 8.1034 470 0.6734 0.5566 0.6734 0.8206
No log 8.1379 472 0.6744 0.5582 0.6744 0.8212
No log 8.1724 474 0.6792 0.5831 0.6792 0.8242
No log 8.2069 476 0.6852 0.5827 0.6852 0.8278
No log 8.2414 478 0.6919 0.5569 0.6919 0.8318
No log 8.2759 480 0.6922 0.5827 0.6922 0.8320
No log 8.3103 482 0.6886 0.5827 0.6886 0.8298
No log 8.3448 484 0.6887 0.5521 0.6887 0.8299
No log 8.3793 486 0.6852 0.5628 0.6852 0.8278
No log 8.4138 488 0.6863 0.5358 0.6863 0.8284
No log 8.4483 490 0.6908 0.5032 0.6908 0.8312
No log 8.4828 492 0.6841 0.5033 0.6841 0.8271
No log 8.5172 494 0.6838 0.5082 0.6838 0.8269
No log 8.5517 496 0.6763 0.5033 0.6763 0.8223
No log 8.5862 498 0.6740 0.5267 0.6740 0.8210
0.3804 8.6207 500 0.6744 0.5295 0.6744 0.8212
0.3804 8.6552 502 0.6752 0.5644 0.6752 0.8217
0.3804 8.6897 504 0.6791 0.5582 0.6791 0.8241
0.3804 8.7241 506 0.6828 0.5733 0.6828 0.8263
0.3804 8.7586 508 0.6811 0.5736 0.6811 0.8253
0.3804 8.7931 510 0.6807 0.5733 0.6807 0.8251
0.3804 8.8276 512 0.6860 0.5629 0.6860 0.8283
0.3804 8.8621 514 0.6919 0.5645 0.6919 0.8318
0.3804 8.8966 516 0.6946 0.5705 0.6946 0.8334
0.3804 8.9310 518 0.6897 0.5629 0.6897 0.8305
0.3804 8.9655 520 0.6864 0.5690 0.6864 0.8285
0.3804 9.0 522 0.6807 0.5736 0.6807 0.8251
0.3804 9.0345 524 0.6790 0.5566 0.6790 0.8240
0.3804 9.0690 526 0.6808 0.5566 0.6808 0.8251
0.3804 9.1034 528 0.6817 0.5567 0.6817 0.8257
0.3804 9.1379 530 0.6853 0.5566 0.6853 0.8278
0.3804 9.1724 532 0.6856 0.5566 0.6856 0.8280
0.3804 9.2069 534 0.6852 0.5566 0.6852 0.8278
0.3804 9.2414 536 0.6851 0.5551 0.6851 0.8277
0.3804 9.2759 538 0.6870 0.5551 0.6870 0.8289
0.3804 9.3103 540 0.6885 0.5551 0.6885 0.8298
0.3804 9.3448 542 0.6923 0.5360 0.6923 0.8321
0.3804 9.3793 544 0.6956 0.5360 0.6956 0.8340
0.3804 9.4138 546 0.6986 0.5551 0.6986 0.8358
0.3804 9.4483 548 0.7011 0.5535 0.7011 0.8373
0.3804 9.4828 550 0.7036 0.5579 0.7036 0.8388
0.3804 9.5172 552 0.7070 0.5358 0.7070 0.8408
0.3804 9.5517 554 0.7117 0.5416 0.7117 0.8436
0.3804 9.5862 556 0.7174 0.5403 0.7174 0.8470
0.3804 9.6207 558 0.7232 0.5390 0.7232 0.8504
0.3804 9.6552 560 0.7270 0.5390 0.7270 0.8526
0.3804 9.6897 562 0.7303 0.5390 0.7303 0.8546
0.3804 9.7241 564 0.7329 0.5390 0.7329 0.8561
0.3804 9.7586 566 0.7353 0.5390 0.7353 0.8575
0.3804 9.7931 568 0.7358 0.5348 0.7358 0.8578
0.3804 9.8276 570 0.7368 0.5348 0.7368 0.8584
0.3804 9.8621 572 0.7368 0.5348 0.7368 0.8584
0.3804 9.8966 574 0.7365 0.5296 0.7365 0.8582
0.3804 9.9310 576 0.7362 0.5296 0.7362 0.8580
0.3804 9.9655 578 0.7358 0.5296 0.7358 0.8578
0.3804 10.0 580 0.7358 0.5296 0.7358 0.8578

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k9_task2_organization

Finetuned
(4023)
this model