ArabicNewSplits5_FineTuningAraBERT_run3_AugV5_k11_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7029
  • Qwk: 0.7003
  • Mse: 0.7029
  • Rmse: 0.8384

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0377 2 2.2625 -0.0045 2.2625 1.5041
No log 0.0755 4 1.6142 0.1256 1.6142 1.2705
No log 0.1132 6 1.4008 0.2183 1.4008 1.1836
No log 0.1509 8 1.3922 0.1771 1.3922 1.1799
No log 0.1887 10 1.3060 0.2515 1.3060 1.1428
No log 0.2264 12 1.3729 0.2786 1.3729 1.1717
No log 0.2642 14 1.3790 0.2406 1.3790 1.1743
No log 0.3019 16 1.2269 0.1764 1.2269 1.1077
No log 0.3396 18 1.1843 0.2153 1.1843 1.0883
No log 0.3774 20 1.1866 0.2913 1.1866 1.0893
No log 0.4151 22 1.1471 0.3040 1.1471 1.0710
No log 0.4528 24 1.1288 0.3198 1.1288 1.0624
No log 0.4906 26 1.1515 0.3399 1.1515 1.0731
No log 0.5283 28 1.1642 0.3537 1.1642 1.0790
No log 0.5660 30 1.2399 0.3444 1.2399 1.1135
No log 0.6038 32 1.4106 0.2242 1.4106 1.1877
No log 0.6415 34 1.4430 0.2137 1.4430 1.2012
No log 0.6792 36 1.2950 0.2259 1.2950 1.1380
No log 0.7170 38 1.2393 0.3161 1.2393 1.1132
No log 0.7547 40 1.1980 0.3119 1.1980 1.0946
No log 0.7925 42 1.1233 0.3917 1.1233 1.0598
No log 0.8302 44 1.0458 0.3887 1.0458 1.0227
No log 0.8679 46 1.0745 0.4064 1.0745 1.0366
No log 0.9057 48 1.1457 0.4028 1.1457 1.0704
No log 0.9434 50 1.0687 0.4473 1.0687 1.0338
No log 0.9811 52 1.1626 0.3608 1.1626 1.0782
No log 1.0189 54 1.4167 0.2516 1.4167 1.1902
No log 1.0566 56 1.3709 0.2761 1.3709 1.1709
No log 1.0943 58 1.1461 0.4004 1.1461 1.0706
No log 1.1321 60 1.0158 0.4342 1.0158 1.0079
No log 1.1698 62 1.3211 0.4147 1.3211 1.1494
No log 1.2075 64 1.4809 0.4217 1.4809 1.2169
No log 1.2453 66 1.2533 0.4262 1.2533 1.1195
No log 1.2830 68 1.0484 0.4422 1.0484 1.0239
No log 1.3208 70 1.0158 0.4401 1.0158 1.0079
No log 1.3585 72 1.0172 0.4401 1.0172 1.0085
No log 1.3962 74 1.0177 0.4584 1.0177 1.0088
No log 1.4340 76 1.0445 0.4705 1.0445 1.0220
No log 1.4717 78 1.2354 0.4463 1.2354 1.1115
No log 1.5094 80 1.5188 0.4139 1.5188 1.2324
No log 1.5472 82 1.3761 0.4067 1.3761 1.1731
No log 1.5849 84 1.0807 0.4779 1.0807 1.0396
No log 1.6226 86 0.8823 0.5654 0.8823 0.9393
No log 1.6604 88 0.8818 0.5541 0.8818 0.9391
No log 1.6981 90 0.8677 0.5541 0.8677 0.9315
No log 1.7358 92 0.8157 0.5595 0.8157 0.9031
No log 1.7736 94 0.8715 0.6182 0.8715 0.9335
No log 1.8113 96 1.0155 0.4812 1.0155 1.0077
No log 1.8491 98 1.0191 0.5189 1.0191 1.0095
No log 1.8868 100 0.9230 0.5591 0.9230 0.9607
No log 1.9245 102 0.8876 0.5812 0.8876 0.9421
No log 1.9623 104 0.8291 0.5997 0.8291 0.9106
No log 2.0 106 0.7881 0.6263 0.7881 0.8878
No log 2.0377 108 0.7856 0.6539 0.7856 0.8863
No log 2.0755 110 0.7978 0.6330 0.7978 0.8932
No log 2.1132 112 0.8891 0.6309 0.8891 0.9429
No log 2.1509 114 1.0972 0.6194 1.0972 1.0475
No log 2.1887 116 1.0868 0.6194 1.0868 1.0425
No log 2.2264 118 0.8737 0.6660 0.8737 0.9347
No log 2.2642 120 0.8426 0.6514 0.8426 0.9179
No log 2.3019 122 0.8804 0.6643 0.8804 0.9383
No log 2.3396 124 0.8763 0.6592 0.8763 0.9361
No log 2.3774 126 0.8628 0.6782 0.8628 0.9288
No log 2.4151 128 0.8641 0.6696 0.8641 0.9296
No log 2.4528 130 0.9231 0.6313 0.9231 0.9608
No log 2.4906 132 0.8509 0.6798 0.8509 0.9225
No log 2.5283 134 0.7971 0.6923 0.7971 0.8928
No log 2.5660 136 0.7568 0.7383 0.7568 0.8699
No log 2.6038 138 0.7184 0.6757 0.7184 0.8476
No log 2.6415 140 0.7190 0.6935 0.7190 0.8479
No log 2.6792 142 0.7376 0.7012 0.7376 0.8589
No log 2.7170 144 0.7752 0.6760 0.7752 0.8805
No log 2.7547 146 0.7349 0.6898 0.7349 0.8573
No log 2.7925 148 0.7033 0.6592 0.7033 0.8386
No log 2.8302 150 0.7029 0.6602 0.7029 0.8384
No log 2.8679 152 0.7064 0.6677 0.7064 0.8405
No log 2.9057 154 0.7846 0.6857 0.7846 0.8858
No log 2.9434 156 0.9703 0.6194 0.9703 0.9850
No log 2.9811 158 1.0541 0.5969 1.0541 1.0267
No log 3.0189 160 1.0682 0.5865 1.0682 1.0335
No log 3.0566 162 0.9773 0.6093 0.9773 0.9886
No log 3.0943 164 0.8829 0.6373 0.8829 0.9396
No log 3.1321 166 0.9082 0.6103 0.9082 0.9530
No log 3.1698 168 0.9694 0.5422 0.9694 0.9846
No log 3.2075 170 0.9488 0.5580 0.9488 0.9740
No log 3.2453 172 0.8324 0.5972 0.8324 0.9124
No log 3.2830 174 0.7622 0.6649 0.7622 0.8731
No log 3.3208 176 0.7670 0.6250 0.7670 0.8758
No log 3.3585 178 0.8003 0.6042 0.8003 0.8946
No log 3.3962 180 0.8483 0.6072 0.8483 0.9210
No log 3.4340 182 0.9064 0.6413 0.9064 0.9520
No log 3.4717 184 0.8440 0.6573 0.8440 0.9187
No log 3.5094 186 0.8485 0.6747 0.8485 0.9211
No log 3.5472 188 0.8438 0.6771 0.8438 0.9186
No log 3.5849 190 0.8154 0.6824 0.8154 0.9030
No log 3.6226 192 0.7579 0.6840 0.7579 0.8706
No log 3.6604 194 0.7568 0.6788 0.7568 0.8699
No log 3.6981 196 0.7667 0.6868 0.7667 0.8756
No log 3.7358 198 0.7428 0.7021 0.7428 0.8619
No log 3.7736 200 0.7129 0.6840 0.7129 0.8443
No log 3.8113 202 0.7348 0.7021 0.7348 0.8572
No log 3.8491 204 0.7974 0.6771 0.7974 0.8930
No log 3.8868 206 0.7539 0.7126 0.7539 0.8683
No log 3.9245 208 0.7227 0.6903 0.7227 0.8501
No log 3.9623 210 0.7112 0.6851 0.7112 0.8433
No log 4.0 212 0.7017 0.6685 0.7017 0.8377
No log 4.0377 214 0.7402 0.6606 0.7402 0.8604
No log 4.0755 216 0.7330 0.6562 0.7330 0.8562
No log 4.1132 218 0.6869 0.6800 0.6869 0.8288
No log 4.1509 220 0.7060 0.7012 0.7060 0.8402
No log 4.1887 222 0.7604 0.7271 0.7604 0.8720
No log 4.2264 224 0.7897 0.7242 0.7897 0.8887
No log 4.2642 226 0.7588 0.7263 0.7588 0.8711
No log 4.3019 228 0.6894 0.6900 0.6894 0.8303
No log 4.3396 230 0.6733 0.6636 0.6733 0.8205
No log 4.3774 232 0.6683 0.6784 0.6683 0.8175
No log 4.4151 234 0.6750 0.6657 0.6750 0.8216
No log 4.4528 236 0.7125 0.6754 0.7125 0.8441
No log 4.4906 238 0.7812 0.6571 0.7812 0.8839
No log 4.5283 240 0.8006 0.6361 0.8006 0.8948
No log 4.5660 242 0.7870 0.6296 0.7870 0.8871
No log 4.6038 244 0.7596 0.6536 0.7596 0.8716
No log 4.6415 246 0.7144 0.6941 0.7144 0.8452
No log 4.6792 248 0.7051 0.7039 0.7051 0.8397
No log 4.7170 250 0.7133 0.7068 0.7133 0.8446
No log 4.7547 252 0.7601 0.7218 0.7601 0.8718
No log 4.7925 254 0.8055 0.7039 0.8055 0.8975
No log 4.8302 256 0.9118 0.6199 0.9118 0.9549
No log 4.8679 258 0.9465 0.6220 0.9465 0.9729
No log 4.9057 260 0.8631 0.6304 0.8631 0.9290
No log 4.9434 262 0.7422 0.7111 0.7422 0.8615
No log 4.9811 264 0.6951 0.7118 0.6951 0.8337
No log 5.0189 266 0.6622 0.7047 0.6622 0.8138
No log 5.0566 268 0.6497 0.6840 0.6497 0.8061
No log 5.0943 270 0.6570 0.6933 0.6570 0.8106
No log 5.1321 272 0.7027 0.7183 0.7027 0.8383
No log 5.1698 274 0.7405 0.6852 0.7405 0.8605
No log 5.2075 276 0.7429 0.6892 0.7429 0.8619
No log 5.2453 278 0.7012 0.6892 0.7012 0.8374
No log 5.2830 280 0.6657 0.6549 0.6657 0.8159
No log 5.3208 282 0.6722 0.6910 0.6722 0.8199
No log 5.3585 284 0.6693 0.6910 0.6693 0.8181
No log 5.3962 286 0.6585 0.7117 0.6585 0.8115
No log 5.4340 288 0.6789 0.6922 0.6789 0.8239
No log 5.4717 290 0.7111 0.7118 0.7111 0.8433
No log 5.5094 292 0.7305 0.7102 0.7305 0.8547
No log 5.5472 294 0.7871 0.6954 0.7871 0.8872
No log 5.5849 296 0.7838 0.6954 0.7838 0.8853
No log 5.6226 298 0.7290 0.7056 0.7290 0.8538
No log 5.6604 300 0.6755 0.7121 0.6755 0.8219
No log 5.6981 302 0.6648 0.7080 0.6648 0.8153
No log 5.7358 304 0.6712 0.6995 0.6712 0.8193
No log 5.7736 306 0.6716 0.6995 0.6716 0.8195
No log 5.8113 308 0.6970 0.7006 0.6970 0.8349
No log 5.8491 310 0.7084 0.7006 0.7084 0.8417
No log 5.8868 312 0.6886 0.7055 0.6886 0.8298
No log 5.9245 314 0.6948 0.7055 0.6948 0.8336
No log 5.9623 316 0.7032 0.7059 0.7032 0.8386
No log 6.0 318 0.6889 0.7059 0.6889 0.8300
No log 6.0377 320 0.6701 0.7139 0.6701 0.8186
No log 6.0755 322 0.6503 0.6949 0.6503 0.8064
No log 6.1132 324 0.6338 0.6927 0.6338 0.7961
No log 6.1509 326 0.6288 0.6907 0.6288 0.7930
No log 6.1887 328 0.6258 0.7037 0.6258 0.7910
No log 6.2264 330 0.6279 0.6978 0.6279 0.7924
No log 6.2642 332 0.6573 0.7269 0.6573 0.8107
No log 6.3019 334 0.6988 0.7216 0.6988 0.8359
No log 6.3396 336 0.6924 0.7216 0.6924 0.8321
No log 6.3774 338 0.6640 0.7269 0.6640 0.8148
No log 6.4151 340 0.6343 0.6818 0.6343 0.7964
No log 6.4528 342 0.6345 0.7081 0.6345 0.7966
No log 6.4906 344 0.6486 0.6475 0.6486 0.8054
No log 6.5283 346 0.6477 0.6598 0.6477 0.8048
No log 6.5660 348 0.6399 0.6763 0.6399 0.7999
No log 6.6038 350 0.6391 0.6905 0.6391 0.7994
No log 6.6415 352 0.6464 0.7046 0.6464 0.8040
No log 6.6792 354 0.6659 0.7194 0.6659 0.8160
No log 6.7170 356 0.6931 0.6855 0.6931 0.8325
No log 6.7547 358 0.6867 0.6961 0.6867 0.8287
No log 6.7925 360 0.6599 0.7198 0.6599 0.8123
No log 6.8302 362 0.6259 0.7260 0.6259 0.7911
No log 6.8679 364 0.6101 0.7048 0.6101 0.7811
No log 6.9057 366 0.6071 0.7122 0.6071 0.7792
No log 6.9434 368 0.6098 0.7090 0.6098 0.7809
No log 6.9811 370 0.6258 0.7260 0.6258 0.7911
No log 7.0189 372 0.6559 0.7229 0.6559 0.8099
No log 7.0566 374 0.6798 0.6990 0.6798 0.8245
No log 7.0943 376 0.7023 0.7075 0.7023 0.8381
No log 7.1321 378 0.6802 0.6990 0.6802 0.8248
No log 7.1698 380 0.6335 0.7277 0.6335 0.7959
No log 7.2075 382 0.6051 0.7208 0.6051 0.7779
No log 7.2453 384 0.5997 0.6939 0.5997 0.7744
No log 7.2830 386 0.6018 0.6948 0.6018 0.7757
No log 7.3208 388 0.6038 0.7024 0.6038 0.7770
No log 7.3585 390 0.6159 0.7341 0.6159 0.7848
No log 7.3962 392 0.6310 0.7351 0.6310 0.7943
No log 7.4340 394 0.6396 0.7227 0.6396 0.7997
No log 7.4717 396 0.6305 0.7267 0.6305 0.7940
No log 7.5094 398 0.6230 0.7178 0.6230 0.7893
No log 7.5472 400 0.6258 0.7178 0.6258 0.7911
No log 7.5849 402 0.6346 0.7178 0.6346 0.7966
No log 7.6226 404 0.6392 0.7222 0.6392 0.7995
No log 7.6604 406 0.6409 0.7074 0.6409 0.8006
No log 7.6981 408 0.6469 0.7147 0.6469 0.8043
No log 7.7358 410 0.6537 0.7168 0.6537 0.8085
No log 7.7736 412 0.6673 0.7023 0.6673 0.8169
No log 7.8113 414 0.6850 0.7067 0.6850 0.8276
No log 7.8491 416 0.7177 0.6825 0.7177 0.8472
No log 7.8868 418 0.7225 0.6864 0.7225 0.8500
No log 7.9245 420 0.7087 0.6912 0.7087 0.8418
No log 7.9623 422 0.7101 0.6912 0.7101 0.8427
No log 8.0 424 0.7100 0.7024 0.7100 0.8426
No log 8.0377 426 0.7190 0.6956 0.7190 0.8479
No log 8.0755 428 0.7105 0.7003 0.7105 0.8429
No log 8.1132 430 0.7078 0.7070 0.7078 0.8413
No log 8.1509 432 0.7129 0.7066 0.7129 0.8443
No log 8.1887 434 0.7295 0.6890 0.7295 0.8541
No log 8.2264 436 0.7463 0.7020 0.7463 0.8639
No log 8.2642 438 0.7688 0.7054 0.7688 0.8768
No log 8.3019 440 0.7723 0.7009 0.7723 0.8788
No log 8.3396 442 0.7554 0.6988 0.7554 0.8692
No log 8.3774 444 0.7226 0.7093 0.7226 0.8501
No log 8.4151 446 0.6873 0.7121 0.6873 0.8290
No log 8.4528 448 0.6585 0.7236 0.6585 0.8115
No log 8.4906 450 0.6472 0.7152 0.6472 0.8045
No log 8.5283 452 0.6426 0.6952 0.6426 0.8016
No log 8.5660 454 0.6467 0.7152 0.6467 0.8042
No log 8.6038 456 0.6576 0.7095 0.6576 0.8109
No log 8.6415 458 0.6807 0.7053 0.6807 0.8250
No log 8.6792 460 0.7088 0.6847 0.7088 0.8419
No log 8.7170 462 0.7220 0.6912 0.7220 0.8497
No log 8.7547 464 0.7192 0.6912 0.7192 0.8480
No log 8.7925 466 0.7088 0.6847 0.7088 0.8419
No log 8.8302 468 0.6949 0.7111 0.6949 0.8336
No log 8.8679 470 0.6798 0.7023 0.6798 0.8245
No log 8.9057 472 0.6735 0.7023 0.6735 0.8207
No log 8.9434 474 0.6664 0.7095 0.6664 0.8163
No log 8.9811 476 0.6596 0.6999 0.6596 0.8122
No log 9.0189 478 0.6532 0.7095 0.6532 0.8082
No log 9.0566 480 0.6483 0.7144 0.6483 0.8052
No log 9.0943 482 0.6452 0.7076 0.6452 0.8032
No log 9.1321 484 0.6436 0.7076 0.6436 0.8023
No log 9.1698 486 0.6456 0.7095 0.6456 0.8035
No log 9.2075 488 0.6524 0.7095 0.6524 0.8077
No log 9.2453 490 0.6617 0.7162 0.6617 0.8134
No log 9.2830 492 0.6740 0.6973 0.6740 0.8210
No log 9.3208 494 0.6821 0.7133 0.6821 0.8259
No log 9.3585 496 0.6864 0.7111 0.6864 0.8285
No log 9.3962 498 0.6899 0.7201 0.6899 0.8306
0.3148 9.4340 500 0.6952 0.7073 0.6952 0.8338
0.3148 9.4717 502 0.6979 0.7073 0.6979 0.8354
0.3148 9.5094 504 0.6984 0.7073 0.6984 0.8357
0.3148 9.5472 506 0.6993 0.7073 0.6993 0.8362
0.3148 9.5849 508 0.7028 0.7052 0.7028 0.8384
0.3148 9.6226 510 0.7085 0.7114 0.7085 0.8417
0.3148 9.6604 512 0.7099 0.7109 0.7099 0.8425
0.3148 9.6981 514 0.7117 0.7109 0.7117 0.8436
0.3148 9.7358 516 0.7134 0.7109 0.7134 0.8446
0.3148 9.7736 518 0.7127 0.7109 0.7127 0.8442
0.3148 9.8113 520 0.7105 0.7109 0.7105 0.8429
0.3148 9.8491 522 0.7085 0.7048 0.7085 0.8417
0.3148 9.8868 524 0.7064 0.7175 0.7064 0.8405
0.3148 9.9245 526 0.7047 0.7003 0.7047 0.8395
0.3148 9.9623 528 0.7036 0.7003 0.7036 0.8388
0.3148 10.0 530 0.7029 0.7003 0.7029 0.8384

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run3_AugV5_k11_task5_organization

Finetuned
(4023)
this model