ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k11_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8784
  • Qwk: 0.4175
  • Mse: 0.8784
  • Rmse: 0.9372

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 4.2093 -0.0087 4.2093 2.0517
No log 0.1379 4 3.0506 -0.0089 3.0506 1.7466
No log 0.2069 6 2.2145 -0.0264 2.2145 1.4881
No log 0.2759 8 1.2738 0.1444 1.2738 1.1286
No log 0.3448 10 1.0849 0.1408 1.0849 1.0416
No log 0.4138 12 1.0723 0.1685 1.0723 1.0355
No log 0.4828 14 1.2675 0.1624 1.2675 1.1258
No log 0.5517 16 1.3089 0.1536 1.3089 1.1441
No log 0.6207 18 1.1215 0.2203 1.1215 1.0590
No log 0.6897 20 0.9812 0.3294 0.9812 0.9906
No log 0.7586 22 1.0172 0.3935 1.0172 1.0086
No log 0.8276 24 1.0019 0.3435 1.0019 1.0009
No log 0.8966 26 0.9642 0.1545 0.9642 0.9819
No log 0.9655 28 0.9517 0.1516 0.9517 0.9756
No log 1.0345 30 0.9777 0.1516 0.9777 0.9888
No log 1.1034 32 0.8983 0.2008 0.8983 0.9478
No log 1.1724 34 0.9063 0.3577 0.9063 0.9520
No log 1.2414 36 1.1562 0.3534 1.1562 1.0753
No log 1.3103 38 1.1708 0.3185 1.1708 1.0820
No log 1.3793 40 0.9924 0.3772 0.9924 0.9962
No log 1.4483 42 0.9773 0.1418 0.9773 0.9886
No log 1.5172 44 1.1067 0.0604 1.1067 1.0520
No log 1.5862 46 1.1339 0.0445 1.1339 1.0648
No log 1.6552 48 1.0950 0.1699 1.0950 1.0464
No log 1.7241 50 1.0634 0.2114 1.0634 1.0312
No log 1.7931 52 1.0246 0.2566 1.0246 1.0122
No log 1.8621 54 0.9945 0.2467 0.9945 0.9973
No log 1.9310 56 0.9659 0.3130 0.9659 0.9828
No log 2.0 58 0.9493 0.3561 0.9493 0.9743
No log 2.0690 60 0.9690 0.2236 0.9690 0.9844
No log 2.1379 62 1.0416 0.1203 1.0416 1.0206
No log 2.2069 64 1.0209 0.1233 1.0209 1.0104
No log 2.2759 66 1.0725 0.0111 1.0725 1.0356
No log 2.3448 68 1.1388 0.1009 1.1388 1.0671
No log 2.4138 70 1.2400 0.0860 1.2400 1.1135
No log 2.4828 72 1.2313 0.0926 1.2313 1.1096
No log 2.5517 74 1.0737 0.0741 1.0737 1.0362
No log 2.6207 76 0.8921 0.3678 0.8921 0.9445
No log 2.6897 78 0.8585 0.3011 0.8585 0.9266
No log 2.7586 80 0.9673 0.2226 0.9673 0.9835
No log 2.8276 82 1.2187 0.1649 1.2187 1.1040
No log 2.8966 84 1.3391 0.1294 1.3391 1.1572
No log 2.9655 86 1.2956 0.0919 1.2956 1.1382
No log 3.0345 88 1.1896 0.1233 1.1896 1.0907
No log 3.1034 90 1.0483 0.2983 1.0483 1.0239
No log 3.1724 92 1.1166 0.3538 1.1166 1.0567
No log 3.2414 94 1.1015 0.3595 1.1015 1.0495
No log 3.3103 96 0.9796 0.3686 0.9796 0.9898
No log 3.3793 98 0.9097 0.375 0.9097 0.9538
No log 3.4483 100 0.7469 0.5307 0.7469 0.8643
No log 3.5172 102 0.6460 0.6154 0.6460 0.8037
No log 3.5862 104 0.6548 0.6324 0.6548 0.8092
No log 3.6552 106 0.9388 0.4752 0.9388 0.9689
No log 3.7241 108 1.2667 0.3581 1.2667 1.1255
No log 3.7931 110 1.3173 0.3505 1.3173 1.1477
No log 3.8621 112 1.1326 0.3739 1.1326 1.0642
No log 3.9310 114 0.8884 0.5318 0.8884 0.9426
No log 4.0 116 0.7433 0.5912 0.7433 0.8622
No log 4.0690 118 0.6256 0.6422 0.6256 0.7909
No log 4.1379 120 0.5898 0.6539 0.5898 0.7680
No log 4.2069 122 0.6117 0.6137 0.6117 0.7821
No log 4.2759 124 0.6351 0.6455 0.6351 0.7969
No log 4.3448 126 0.6087 0.6849 0.6087 0.7802
No log 4.4138 128 0.5890 0.6958 0.5890 0.7674
No log 4.4828 130 0.5608 0.6996 0.5608 0.7489
No log 4.5517 132 0.5493 0.6715 0.5493 0.7412
No log 4.6207 134 0.5912 0.6588 0.5912 0.7689
No log 4.6897 136 0.7566 0.6291 0.7566 0.8698
No log 4.7586 138 0.9292 0.4969 0.9292 0.9640
No log 4.8276 140 0.9416 0.5321 0.9416 0.9703
No log 4.8966 142 0.7637 0.5279 0.7637 0.8739
No log 4.9655 144 0.7237 0.5416 0.7237 0.8507
No log 5.0345 146 0.7872 0.5579 0.7872 0.8872
No log 5.1034 148 0.9754 0.4885 0.9754 0.9876
No log 5.1724 150 1.1185 0.3544 1.1185 1.0576
No log 5.2414 152 0.9296 0.5078 0.9296 0.9642
No log 5.3103 154 0.6410 0.5400 0.6410 0.8006
No log 5.3793 156 0.6198 0.5568 0.6198 0.7873
No log 5.4483 158 0.6643 0.5614 0.6643 0.8151
No log 5.5172 160 0.5872 0.5316 0.5872 0.7663
No log 5.5862 162 0.6022 0.5927 0.6022 0.7760
No log 5.6552 164 0.7286 0.5867 0.7286 0.8536
No log 5.7241 166 0.8821 0.5908 0.8821 0.9392
No log 5.7931 168 0.8585 0.5644 0.8585 0.9266
No log 5.8621 170 0.7845 0.5561 0.7845 0.8857
No log 5.9310 172 0.7900 0.5561 0.7900 0.8888
No log 6.0 174 0.8020 0.5994 0.8020 0.8955
No log 6.0690 176 0.6995 0.5618 0.6995 0.8364
No log 6.1379 178 0.6019 0.6327 0.6019 0.7758
No log 6.2069 180 0.5716 0.5823 0.5716 0.7561
No log 6.2759 182 0.5831 0.5823 0.5831 0.7636
No log 6.3448 184 0.6070 0.6255 0.6070 0.7791
No log 6.4138 186 0.6538 0.5981 0.6538 0.8086
No log 6.4828 188 0.6934 0.5763 0.6934 0.8327
No log 6.5517 190 0.6524 0.5546 0.6524 0.8077
No log 6.6207 192 0.6428 0.5546 0.6428 0.8017
No log 6.6897 194 0.6163 0.6051 0.6163 0.7850
No log 6.7586 196 0.6061 0.5168 0.6061 0.7785
No log 6.8276 198 0.6089 0.5142 0.6089 0.7803
No log 6.8966 200 0.6764 0.6039 0.6764 0.8224
No log 6.9655 202 0.8600 0.4987 0.8600 0.9274
No log 7.0345 204 0.9305 0.4975 0.9305 0.9646
No log 7.1034 206 0.8925 0.4767 0.8925 0.9447
No log 7.1724 208 0.7457 0.5400 0.7457 0.8636
No log 7.2414 210 0.6852 0.5490 0.6852 0.8278
No log 7.3103 212 0.6971 0.5966 0.6971 0.8349
No log 7.3793 214 0.7197 0.5463 0.7197 0.8483
No log 7.4483 216 0.7906 0.4932 0.7906 0.8892
No log 7.5172 218 0.8324 0.5154 0.8324 0.9123
No log 7.5862 220 0.8197 0.4926 0.8197 0.9054
No log 7.6552 222 0.8210 0.4926 0.8210 0.9061
No log 7.7241 224 0.8118 0.4186 0.8118 0.9010
No log 7.7931 226 0.7599 0.4745 0.7599 0.8717
No log 7.8621 228 0.7401 0.5127 0.7401 0.8603
No log 7.9310 230 0.7029 0.4642 0.7029 0.8384
No log 8.0 232 0.6666 0.5142 0.6666 0.8164
No log 8.0690 234 0.6514 0.6154 0.6514 0.8071
No log 8.1379 236 0.6879 0.6300 0.6879 0.8294
No log 8.2069 238 0.7414 0.6442 0.7414 0.8611
No log 8.2759 240 0.7176 0.6442 0.7176 0.8471
No log 8.3448 242 0.5946 0.6906 0.5946 0.7711
No log 8.4138 244 0.5486 0.6219 0.5486 0.7407
No log 8.4828 246 0.5302 0.6286 0.5302 0.7281
No log 8.5517 248 0.5321 0.6425 0.5321 0.7294
No log 8.6207 250 0.5587 0.6425 0.5587 0.7474
No log 8.6897 252 0.5888 0.6043 0.5888 0.7673
No log 8.7586 254 0.6476 0.6740 0.6476 0.8048
No log 8.8276 256 0.6799 0.6240 0.6799 0.8246
No log 8.8966 258 0.6889 0.5948 0.6889 0.8300
No log 8.9655 260 0.6675 0.4643 0.6675 0.8170
No log 9.0345 262 0.6681 0.4295 0.6681 0.8174
No log 9.1034 264 0.6458 0.5359 0.6458 0.8036
No log 9.1724 266 0.6352 0.5577 0.6352 0.7970
No log 9.2414 268 0.6272 0.5010 0.6272 0.7920
No log 9.3103 270 0.6685 0.5645 0.6685 0.8176
No log 9.3793 272 0.7205 0.5912 0.7205 0.8488
No log 9.4483 274 0.6921 0.6240 0.6921 0.8319
No log 9.5172 276 0.6387 0.5798 0.6387 0.7992
No log 9.5862 278 0.5986 0.5259 0.5986 0.7737
No log 9.6552 280 0.5777 0.5259 0.5777 0.7601
No log 9.7241 282 0.5661 0.5960 0.5661 0.7524
No log 9.7931 284 0.5727 0.6014 0.5727 0.7568
No log 9.8621 286 0.5989 0.6305 0.5989 0.7739
No log 9.9310 288 0.6152 0.6491 0.6152 0.7844
No log 10.0 290 0.6233 0.6167 0.6233 0.7895
No log 10.0690 292 0.6327 0.6352 0.6327 0.7954
No log 10.1379 294 0.6441 0.6167 0.6441 0.8025
No log 10.2069 296 0.6647 0.6053 0.6647 0.8153
No log 10.2759 298 0.6355 0.5259 0.6355 0.7972
No log 10.3448 300 0.6187 0.4498 0.6187 0.7866
No log 10.4138 302 0.6137 0.5018 0.6137 0.7834
No log 10.4828 304 0.6159 0.5377 0.6159 0.7848
No log 10.5517 306 0.6425 0.6529 0.6425 0.8016
No log 10.6207 308 0.6421 0.6143 0.6421 0.8013
No log 10.6897 310 0.6729 0.6157 0.6729 0.8203
No log 10.7586 312 0.7406 0.6089 0.7406 0.8606
No log 10.8276 314 0.7337 0.6089 0.7337 0.8565
No log 10.8966 316 0.6910 0.5912 0.6910 0.8313
No log 10.9655 318 0.6770 0.6127 0.6770 0.8228
No log 11.0345 320 0.6371 0.6773 0.6371 0.7982
No log 11.1034 322 0.5923 0.6415 0.5923 0.7696
No log 11.1724 324 0.6029 0.6210 0.6029 0.7764
No log 11.2414 326 0.6178 0.5850 0.6178 0.7860
No log 11.3103 328 0.6433 0.5823 0.6433 0.8021
No log 11.3793 330 0.6884 0.6167 0.6884 0.8297
No log 11.4483 332 0.7244 0.5822 0.7244 0.8511
No log 11.5172 334 0.8236 0.5346 0.8236 0.9075
No log 11.5862 336 0.9530 0.4754 0.9530 0.9762
No log 11.6552 338 0.9133 0.4871 0.9133 0.9557
No log 11.7241 340 0.7661 0.5157 0.7661 0.8753
No log 11.7931 342 0.6697 0.5850 0.6697 0.8184
No log 11.8621 344 0.6773 0.5377 0.6773 0.8230
No log 11.9310 346 0.7329 0.5534 0.7329 0.8561
No log 12.0 348 0.8736 0.4993 0.8736 0.9347
No log 12.0690 350 0.9324 0.4754 0.9324 0.9656
No log 12.1379 352 0.8461 0.4771 0.8461 0.9198
No log 12.2069 354 0.7320 0.5059 0.7320 0.8556
No log 12.2759 356 0.6876 0.5698 0.6876 0.8292
No log 12.3448 358 0.7247 0.4836 0.7247 0.8513
No log 12.4138 360 0.8260 0.5344 0.8260 0.9089
No log 12.4828 362 0.8957 0.4767 0.8957 0.9464
No log 12.5517 364 0.8325 0.5451 0.8325 0.9124
No log 12.6207 366 0.7159 0.5318 0.7159 0.8461
No log 12.6897 368 0.6753 0.5721 0.6753 0.8217
No log 12.7586 370 0.6810 0.5850 0.6810 0.8252
No log 12.8276 372 0.7370 0.5602 0.7370 0.8585
No log 12.8966 374 0.7710 0.5602 0.7710 0.8781
No log 12.9655 376 0.7104 0.5718 0.7104 0.8429
No log 13.0345 378 0.6794 0.5823 0.6794 0.8243
No log 13.1034 380 0.6624 0.5978 0.6624 0.8139
No log 13.1724 382 0.6596 0.6282 0.6596 0.8121
No log 13.2414 384 0.6535 0.6491 0.6535 0.8084
No log 13.3103 386 0.6263 0.6025 0.6263 0.7914
No log 13.3793 388 0.6191 0.6054 0.6191 0.7868
No log 13.4483 390 0.6285 0.5498 0.6285 0.7928
No log 13.5172 392 0.6407 0.5516 0.6407 0.8004
No log 13.5862 394 0.6589 0.5585 0.6589 0.8117
No log 13.6552 396 0.7238 0.6178 0.7238 0.8508
No log 13.7241 398 0.7678 0.5695 0.7678 0.8762
No log 13.7931 400 0.8091 0.5579 0.8091 0.8995
No log 13.8621 402 0.8633 0.5319 0.8633 0.9291
No log 13.9310 404 0.8567 0.5319 0.8567 0.9256
No log 14.0 406 0.8021 0.5370 0.8021 0.8956
No log 14.0690 408 0.7883 0.5266 0.7883 0.8879
No log 14.1379 410 0.7488 0.5397 0.7488 0.8653
No log 14.2069 412 0.7344 0.5397 0.7344 0.8569
No log 14.2759 414 0.7702 0.5927 0.7702 0.8776
No log 14.3448 416 0.8585 0.5098 0.8585 0.9266
No log 14.4138 418 0.9069 0.4415 0.9069 0.9523
No log 14.4828 420 0.8392 0.5107 0.8392 0.9161
No log 14.5517 422 0.7243 0.5916 0.7243 0.8511
No log 14.6207 424 0.6404 0.6499 0.6404 0.8002
No log 14.6897 426 0.6285 0.6697 0.6285 0.7928
No log 14.7586 428 0.6590 0.5846 0.6590 0.8118
No log 14.8276 430 0.7651 0.5684 0.7651 0.8747
No log 14.8966 432 0.8382 0.5102 0.8382 0.9155
No log 14.9655 434 0.7885 0.5451 0.7885 0.8880
No log 15.0345 436 0.6878 0.5534 0.6878 0.8294
No log 15.1034 438 0.6775 0.5331 0.6775 0.8231
No log 15.1724 440 0.6922 0.5331 0.6922 0.8320
No log 15.2414 442 0.7234 0.5397 0.7234 0.8506
No log 15.3103 444 0.7738 0.5266 0.7738 0.8797
No log 15.3793 446 0.7545 0.5697 0.7545 0.8686
No log 15.4483 448 0.7040 0.5898 0.7040 0.8390
No log 15.5172 450 0.6636 0.5763 0.6636 0.8146
No log 15.5862 452 0.6518 0.6133 0.6518 0.8074
No log 15.6552 454 0.6607 0.6133 0.6607 0.8128
No log 15.7241 456 0.6929 0.6035 0.6929 0.8324
No log 15.7931 458 0.7883 0.5799 0.7883 0.8879
No log 15.8621 460 0.8350 0.5891 0.8350 0.9138
No log 15.9310 462 0.7752 0.5695 0.7752 0.8804
No log 16.0 464 0.7709 0.5497 0.7709 0.8780
No log 16.0690 466 0.8072 0.5479 0.8072 0.8985
No log 16.1379 468 0.8686 0.5194 0.8686 0.9320
No log 16.2069 470 0.8624 0.5089 0.8624 0.9287
No log 16.2759 472 0.8185 0.5372 0.8185 0.9047
No log 16.3448 474 0.8358 0.4769 0.8358 0.9142
No log 16.4138 476 0.8109 0.5447 0.8109 0.9005
No log 16.4828 478 0.7439 0.5695 0.7439 0.8625
No log 16.5517 480 0.6847 0.6272 0.6847 0.8275
No log 16.6207 482 0.6849 0.5948 0.6849 0.8276
No log 16.6897 484 0.6922 0.6343 0.6922 0.8320
No log 16.7586 486 0.7315 0.5988 0.7315 0.8553
No log 16.8276 488 0.7907 0.5332 0.7907 0.8892
No log 16.8966 490 0.7836 0.5332 0.7836 0.8852
No log 16.9655 492 0.7051 0.5981 0.7051 0.8397
No log 17.0345 494 0.6950 0.6071 0.6950 0.8337
No log 17.1034 496 0.7640 0.5447 0.7640 0.8741
No log 17.1724 498 0.8999 0.4873 0.8999 0.9487
0.3012 17.2414 500 1.0566 0.4746 1.0566 1.0279
0.3012 17.3103 502 1.0578 0.4747 1.0578 1.0285
0.3012 17.3793 504 0.9254 0.4655 0.9254 0.9620
0.3012 17.4483 506 0.8298 0.5463 0.8298 0.9109
0.3012 17.5172 508 0.7765 0.5292 0.7765 0.8812
0.3012 17.5862 510 0.7582 0.4579 0.7582 0.8708
0.3012 17.6552 512 0.7741 0.4579 0.7741 0.8798
0.3012 17.7241 514 0.8509 0.4666 0.8509 0.9224
0.3012 17.7931 516 0.9054 0.3830 0.9054 0.9515
0.3012 17.8621 518 0.8784 0.4175 0.8784 0.9372

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k11_task5_organization

Finetuned
(4023)
this model