ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k1_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8347
  • Qwk: 0.6809
  • Mse: 0.8347
  • Rmse: 0.9136

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.25 2 8.6832 0.0170 8.6832 2.9467
No log 0.5 4 5.8573 0.0453 5.8573 2.4202
No log 0.75 6 3.0685 0.1562 3.0685 1.7517
No log 1.0 8 2.0276 0.2154 2.0276 1.4239
No log 1.25 10 1.4738 0.2478 1.4738 1.2140
No log 1.5 12 1.5619 0.2807 1.5619 1.2498
No log 1.75 14 1.5118 0.2478 1.5118 1.2295
No log 2.0 16 1.6008 0.2478 1.6008 1.2652
No log 2.25 18 1.8900 0.2787 1.8900 1.3748
No log 2.5 20 2.1897 0.1550 2.1897 1.4798
No log 2.75 22 1.6879 0.2759 1.6879 1.2992
No log 3.0 24 1.3683 0.2523 1.3683 1.1697
No log 3.25 26 1.3144 0.2703 1.3144 1.1465
No log 3.5 28 1.2772 0.4500 1.2772 1.1301
No log 3.75 30 1.2653 0.4202 1.2653 1.1249
No log 4.0 32 1.7177 0.3306 1.7177 1.3106
No log 4.25 34 2.0059 0.3030 2.0059 1.4163
No log 4.5 36 1.3115 0.3279 1.3115 1.1452
No log 4.75 38 1.0048 0.5891 1.0048 1.0024
No log 5.0 40 1.3274 0.4355 1.3274 1.1521
No log 5.25 42 1.4822 0.35 1.4822 1.2174
No log 5.5 44 1.2618 0.5414 1.2618 1.1233
No log 5.75 46 0.9645 0.6260 0.9645 0.9821
No log 6.0 48 1.1281 0.4444 1.1281 1.0621
No log 6.25 50 1.6603 0.4173 1.6603 1.2885
No log 6.5 52 1.2966 0.3780 1.2966 1.1387
No log 6.75 54 0.8803 0.6412 0.8803 0.9383
No log 7.0 56 0.8341 0.6515 0.8341 0.9133
No log 7.25 58 0.8120 0.6047 0.8120 0.9011
No log 7.5 60 0.8317 0.6471 0.8317 0.9120
No log 7.75 62 0.8729 0.6906 0.8729 0.9343
No log 8.0 64 0.9211 0.6466 0.9211 0.9597
No log 8.25 66 0.9172 0.6418 0.9172 0.9577
No log 8.5 68 0.7916 0.6815 0.7916 0.8897
No log 8.75 70 0.7010 0.7206 0.7010 0.8373
No log 9.0 72 0.7389 0.6906 0.7389 0.8596
No log 9.25 74 0.7528 0.7 0.7528 0.8676
No log 9.5 76 0.7530 0.7 0.7530 0.8678
No log 9.75 78 0.6969 0.7299 0.6969 0.8348
No log 10.0 80 0.7101 0.7299 0.7101 0.8427
No log 10.25 82 0.7305 0.7059 0.7305 0.8547
No log 10.5 84 0.7541 0.7059 0.7541 0.8684
No log 10.75 86 0.7894 0.7059 0.7894 0.8885
No log 11.0 88 0.7990 0.7153 0.7990 0.8939
No log 11.25 90 0.9162 0.5672 0.9162 0.9572
No log 11.5 92 0.8418 0.6418 0.8418 0.9175
No log 11.75 94 0.7508 0.6912 0.7508 0.8665
No log 12.0 96 0.7619 0.7194 0.7619 0.8729
No log 12.25 98 0.7788 0.7234 0.7788 0.8825
No log 12.5 100 0.7314 0.6815 0.7314 0.8552
No log 12.75 102 0.7325 0.7092 0.7325 0.8558
No log 13.0 104 0.7511 0.7407 0.7511 0.8666
No log 13.25 106 0.7182 0.7421 0.7182 0.8475
No log 13.5 108 0.7330 0.7578 0.7330 0.8561
No log 13.75 110 0.8123 0.7296 0.8123 0.9013
No log 14.0 112 0.8663 0.7006 0.8663 0.9308
No log 14.25 114 0.8062 0.6667 0.8062 0.8979
No log 14.5 116 0.7675 0.6562 0.7675 0.8761
No log 14.75 118 0.6881 0.7059 0.6881 0.8295
No log 15.0 120 0.6419 0.7606 0.6419 0.8012
No log 15.25 122 0.7743 0.6714 0.7743 0.8800
No log 15.5 124 0.8546 0.6522 0.8546 0.9244
No log 15.75 126 0.7008 0.7353 0.7008 0.8372
No log 16.0 128 0.8763 0.6909 0.8763 0.9361
No log 16.25 130 1.0305 0.6667 1.0305 1.0152
No log 16.5 132 0.7718 0.7160 0.7718 0.8785
No log 16.75 134 0.6343 0.7606 0.6343 0.7964
No log 17.0 136 0.7468 0.7172 0.7468 0.8642
No log 17.25 138 0.6777 0.7042 0.6777 0.8232
No log 17.5 140 0.6930 0.7237 0.6930 0.8324
No log 17.75 142 0.8728 0.6982 0.8728 0.9343
No log 18.0 144 0.8995 0.6982 0.8995 0.9484
No log 18.25 146 0.7673 0.7160 0.7673 0.8760
No log 18.5 148 0.6629 0.7582 0.6629 0.8142
No log 18.75 150 0.6637 0.7297 0.6637 0.8147
No log 19.0 152 0.7162 0.7092 0.7162 0.8463
No log 19.25 154 0.8105 0.6901 0.8105 0.9003
No log 19.5 156 0.9469 0.6755 0.9469 0.9731
No log 19.75 158 1.0775 0.6220 1.0775 1.0380
No log 20.0 160 1.0099 0.6364 1.0099 1.0050
No log 20.25 162 0.8222 0.6957 0.8222 0.9067
No log 20.5 164 0.7115 0.7164 0.7115 0.8435
No log 20.75 166 0.6747 0.7194 0.6747 0.8214
No log 21.0 168 0.6491 0.7778 0.6491 0.8056
No log 21.25 170 0.7449 0.7215 0.7449 0.8631
No log 21.5 172 1.0263 0.6667 1.0263 1.0130
No log 21.75 174 1.1461 0.6012 1.1461 1.0706
No log 22.0 176 1.0632 0.6494 1.0632 1.0311
No log 22.25 178 0.9718 0.6143 0.9718 0.9858
No log 22.5 180 0.8898 0.6331 0.8898 0.9433
No log 22.75 182 0.8441 0.6475 0.8441 0.9188
No log 23.0 184 0.8218 0.6528 0.8218 0.9065
No log 23.25 186 0.8231 0.6901 0.8231 0.9073
No log 23.5 188 0.8313 0.6901 0.8313 0.9118
No log 23.75 190 0.8272 0.6812 0.8272 0.9095
No log 24.0 192 0.8327 0.6715 0.8327 0.9125
No log 24.25 194 0.8562 0.6522 0.8562 0.9253
No log 24.5 196 0.8479 0.6667 0.8479 0.9208
No log 24.75 198 0.8124 0.7152 0.8124 0.9013
No log 25.0 200 0.7182 0.7483 0.7182 0.8475
No log 25.25 202 0.6869 0.7483 0.6869 0.8288
No log 25.5 204 0.6966 0.7333 0.6966 0.8346
No log 25.75 206 0.7407 0.7389 0.7407 0.8606
No log 26.0 208 0.7442 0.7515 0.7442 0.8627
No log 26.25 210 0.7085 0.7467 0.7085 0.8417
No log 26.5 212 0.7136 0.7467 0.7136 0.8447
No log 26.75 214 0.7062 0.7517 0.7062 0.8404
No log 27.0 216 0.7173 0.7432 0.7173 0.8470
No log 27.25 218 0.7581 0.7089 0.7581 0.8707
No log 27.5 220 0.8196 0.7205 0.8196 0.9053
No log 27.75 222 0.7978 0.6897 0.7978 0.8932
No log 28.0 224 0.8244 0.6980 0.8244 0.9080
No log 28.25 226 0.8227 0.6897 0.8227 0.9070
No log 28.5 228 0.8421 0.6761 0.8421 0.9177
No log 28.75 230 0.8699 0.6761 0.8699 0.9327
No log 29.0 232 0.8249 0.6897 0.8249 0.9082
No log 29.25 234 0.8193 0.6897 0.8193 0.9051
No log 29.5 236 0.7864 0.7027 0.7864 0.8868
No log 29.75 238 0.7779 0.7027 0.7779 0.8820
No log 30.0 240 0.7652 0.7114 0.7652 0.8747
No log 30.25 242 0.7354 0.7285 0.7354 0.8576
No log 30.5 244 0.7462 0.7114 0.7463 0.8639
No log 30.75 246 0.7521 0.7114 0.7521 0.8672
No log 31.0 248 0.7421 0.6950 0.7421 0.8615
No log 31.25 250 0.7568 0.6950 0.7568 0.8699
No log 31.5 252 0.7822 0.6761 0.7822 0.8844
No log 31.75 254 0.7839 0.6897 0.7839 0.8854
No log 32.0 256 0.8089 0.6849 0.8089 0.8994
No log 32.25 258 0.7671 0.7027 0.7671 0.8758
No log 32.5 260 0.7190 0.7067 0.7190 0.8479
No log 32.75 262 0.6971 0.7703 0.6971 0.8349
No log 33.0 264 0.7055 0.7724 0.7055 0.8399
No log 33.25 266 0.7371 0.7483 0.7371 0.8585
No log 33.5 268 0.8228 0.6713 0.8228 0.9071
No log 33.75 270 0.9641 0.6543 0.9641 0.9819
No log 34.0 272 0.9480 0.6667 0.9480 0.9737
No log 34.25 274 0.8256 0.6857 0.8256 0.9086
No log 34.5 276 0.7940 0.7206 0.7940 0.8910
No log 34.75 278 0.7977 0.6970 0.7977 0.8931
No log 35.0 280 0.8209 0.7111 0.8209 0.9061
No log 35.25 282 0.8292 0.6715 0.8292 0.9106
No log 35.5 284 0.8073 0.6713 0.8073 0.8985
No log 35.75 286 0.7903 0.6912 0.7903 0.8890
No log 36.0 288 0.7781 0.7143 0.7781 0.8821
No log 36.25 290 0.7588 0.7234 0.7588 0.8711
No log 36.5 292 0.7274 0.7361 0.7274 0.8529
No log 36.75 294 0.6970 0.7273 0.6970 0.8349
No log 37.0 296 0.6905 0.7324 0.6905 0.8310
No log 37.25 298 0.6987 0.7518 0.6987 0.8359
No log 37.5 300 0.7063 0.7518 0.7063 0.8404
No log 37.75 302 0.7291 0.7286 0.7291 0.8539
No log 38.0 304 0.7418 0.7050 0.7418 0.8613
No log 38.25 306 0.7513 0.7050 0.7513 0.8668
No log 38.5 308 0.8230 0.6974 0.8230 0.9072
No log 38.75 310 0.8655 0.7081 0.8655 0.9303
No log 39.0 312 0.8581 0.7081 0.8581 0.9263
No log 39.25 314 0.8014 0.6968 0.8014 0.8952
No log 39.5 316 0.7343 0.7183 0.7343 0.8569
No log 39.75 318 0.7008 0.75 0.7008 0.8371
No log 40.0 320 0.7029 0.75 0.7029 0.8384
No log 40.25 322 0.7105 0.7317 0.7105 0.8429
No log 40.5 324 0.6907 0.7190 0.6907 0.8311
No log 40.75 326 0.6858 0.7483 0.6858 0.8281
No log 41.0 328 0.6809 0.7671 0.6809 0.8251
No log 41.25 330 0.7027 0.7376 0.7027 0.8382
No log 41.5 332 0.7261 0.7111 0.7261 0.8521
No log 41.75 334 0.7507 0.7429 0.7507 0.8665
No log 42.0 336 0.8136 0.7034 0.8136 0.9020
No log 42.25 338 0.8680 0.6974 0.8680 0.9316
No log 42.5 340 0.8322 0.6980 0.8322 0.9122
No log 42.75 342 0.7684 0.7211 0.7684 0.8766
No log 43.0 344 0.7090 0.7639 0.7090 0.8420
No log 43.25 346 0.6825 0.7919 0.6825 0.8261
No log 43.5 348 0.6847 0.8079 0.6847 0.8275
No log 43.75 350 0.6912 0.8079 0.6912 0.8314
No log 44.0 352 0.7263 0.7432 0.7263 0.8522
No log 44.25 354 0.7644 0.7361 0.7644 0.8743
No log 44.5 356 0.7892 0.7007 0.7892 0.8884
No log 44.75 358 0.8070 0.6357 0.8070 0.8983
No log 45.0 360 0.8088 0.6562 0.8088 0.8993
No log 45.25 362 0.8029 0.6617 0.8029 0.8961
No log 45.5 364 0.7775 0.7042 0.7775 0.8818
No log 45.75 366 0.7239 0.7234 0.7239 0.8509
No log 46.0 368 0.6924 0.8 0.6924 0.8321
No log 46.25 370 0.6881 0.7891 0.6881 0.8295
No log 46.5 372 0.6856 0.7703 0.6856 0.8280
No log 46.75 374 0.6758 0.7973 0.6758 0.8221
No log 47.0 376 0.6758 0.7867 0.6758 0.8221
No log 47.25 378 0.6843 0.7785 0.6843 0.8272
No log 47.5 380 0.6820 0.7785 0.6820 0.8258
No log 47.75 382 0.6682 0.7867 0.6682 0.8175
No log 48.0 384 0.6760 0.7671 0.6760 0.8222
No log 48.25 386 0.6947 0.7671 0.6947 0.8335
No log 48.5 388 0.7112 0.7059 0.7112 0.8433
No log 48.75 390 0.7552 0.7 0.7552 0.8690
No log 49.0 392 0.7859 0.6806 0.7859 0.8865
No log 49.25 394 0.7868 0.6806 0.7868 0.8870
No log 49.5 396 0.7973 0.6806 0.7973 0.8929
No log 49.75 398 0.8115 0.6950 0.8115 0.9008
No log 50.0 400 0.8586 0.6857 0.8586 0.9266
No log 50.25 402 0.8894 0.6857 0.8894 0.9431
No log 50.5 404 0.9180 0.6939 0.9180 0.9581
No log 50.75 406 0.8879 0.7083 0.8879 0.9423
No log 51.0 408 0.8155 0.6993 0.8155 0.9030
No log 51.25 410 0.7274 0.7194 0.7274 0.8529
No log 51.5 412 0.6882 0.7482 0.6882 0.8296
No log 51.75 414 0.6628 0.7714 0.6628 0.8141
No log 52.0 416 0.6553 0.7714 0.6552 0.8095
No log 52.25 418 0.6511 0.7801 0.6511 0.8069
No log 52.5 420 0.6515 0.7482 0.6515 0.8072
No log 52.75 422 0.6902 0.6806 0.6902 0.8308
No log 53.0 424 0.7459 0.6846 0.7459 0.8637
No log 53.25 426 0.7428 0.6842 0.7428 0.8619
No log 53.5 428 0.7212 0.6993 0.7212 0.8492
No log 53.75 430 0.7280 0.7246 0.7280 0.8532
No log 54.0 432 0.7471 0.7068 0.7471 0.8644
No log 54.25 434 0.7680 0.7068 0.7680 0.8764
No log 54.5 436 0.7967 0.7092 0.7967 0.8926
No log 54.75 438 0.8097 0.6809 0.8097 0.8998
No log 55.0 440 0.8621 0.6713 0.8621 0.9285
No log 55.25 442 0.8897 0.6622 0.8897 0.9432
No log 55.5 444 0.8718 0.6620 0.8718 0.9337
No log 55.75 446 0.8341 0.6809 0.8341 0.9133
No log 56.0 448 0.8040 0.7111 0.8040 0.8967
No log 56.25 450 0.7940 0.7111 0.7940 0.8911
No log 56.5 452 0.8007 0.7111 0.8007 0.8948
No log 56.75 454 0.8300 0.6763 0.8300 0.9110
No log 57.0 456 0.8419 0.6571 0.8419 0.9176
No log 57.25 458 0.8627 0.6571 0.8627 0.9288
No log 57.5 460 0.8510 0.6667 0.8510 0.9225
No log 57.75 462 0.8409 0.6715 0.8409 0.9170
No log 58.0 464 0.8429 0.6912 0.8429 0.9181
No log 58.25 466 0.8280 0.6912 0.8280 0.9099
No log 58.5 468 0.8121 0.6912 0.8121 0.9012
No log 58.75 470 0.8130 0.6812 0.8130 0.9017
No log 59.0 472 0.8135 0.6667 0.8135 0.9020
No log 59.25 474 0.8072 0.6933 0.8072 0.8985
No log 59.5 476 0.7644 0.7059 0.7644 0.8743
No log 59.75 478 0.7154 0.6993 0.7154 0.8458
No log 60.0 480 0.7083 0.7092 0.7083 0.8416
No log 60.25 482 0.7243 0.7246 0.7243 0.8511
No log 60.5 484 0.7517 0.6857 0.7517 0.8670
No log 60.75 486 0.7750 0.6857 0.7750 0.8803
No log 61.0 488 0.7822 0.6912 0.7822 0.8844
No log 61.25 490 0.7684 0.6912 0.7684 0.8766
No log 61.5 492 0.7450 0.6912 0.7450 0.8631
No log 61.75 494 0.7514 0.6912 0.7514 0.8668
No log 62.0 496 0.7521 0.6912 0.7521 0.8672
No log 62.25 498 0.7424 0.6912 0.7424 0.8616
0.3031 62.5 500 0.7169 0.6901 0.7169 0.8467
0.3031 62.75 502 0.6923 0.7143 0.6923 0.8321
0.3031 63.0 504 0.6952 0.6901 0.6952 0.8338
0.3031 63.25 506 0.7087 0.7034 0.7087 0.8418
0.3031 63.5 508 0.7190 0.6901 0.7190 0.8480
0.3031 63.75 510 0.7205 0.7050 0.7205 0.8488
0.3031 64.0 512 0.7368 0.6912 0.7368 0.8583
0.3031 64.25 514 0.7592 0.6901 0.7592 0.8713
0.3031 64.5 516 0.7961 0.6806 0.7961 0.8922
0.3031 64.75 518 0.7948 0.6806 0.7948 0.8915
0.3031 65.0 520 0.7782 0.6763 0.7782 0.8821
0.3031 65.25 522 0.7715 0.6912 0.7715 0.8783
0.3031 65.5 524 0.7651 0.7111 0.7651 0.8747
0.3031 65.75 526 0.7538 0.7111 0.7538 0.8682
0.3031 66.0 528 0.7484 0.7206 0.7484 0.8651
0.3031 66.25 530 0.7658 0.6763 0.7658 0.8751
0.3031 66.5 532 0.8074 0.6713 0.8074 0.8985
0.3031 66.75 534 0.8306 0.6933 0.8306 0.9114
0.3031 67.0 536 0.8159 0.6806 0.8159 0.9033
0.3031 67.25 538 0.7825 0.6763 0.7825 0.8846
0.3031 67.5 540 0.7668 0.7206 0.7668 0.8757
0.3031 67.75 542 0.7765 0.7206 0.7765 0.8812
0.3031 68.0 544 0.8004 0.6912 0.8004 0.8947
0.3031 68.25 546 0.8184 0.6763 0.8184 0.9046
0.3031 68.5 548 0.8244 0.6912 0.8244 0.9079
0.3031 68.75 550 0.8397 0.6763 0.8397 0.9164
0.3031 69.0 552 0.8531 0.6667 0.8531 0.9236
0.3031 69.25 554 0.8773 0.6713 0.8773 0.9366
0.3031 69.5 556 0.8685 0.6620 0.8685 0.9319
0.3031 69.75 558 0.8347 0.6809 0.8347 0.9136

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run1_AugV5_k1_task1_organization

Finetuned
(4023)
this model