ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k12_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8019
  • Qwk: 0.6912
  • Mse: 0.8019
  • Rmse: 0.8955

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0357 2 7.1597 0.0056 7.1597 2.6758
No log 0.0714 4 4.9340 0.0157 4.9340 2.2213
No log 0.1071 6 4.0670 -0.0773 4.0670 2.0167
No log 0.1429 8 2.5122 0.1351 2.5122 1.5850
No log 0.1786 10 1.9094 0.2478 1.9094 1.3818
No log 0.2143 12 1.8990 0.2414 1.8990 1.3780
No log 0.25 14 2.0481 0.2131 2.0481 1.4311
No log 0.2857 16 1.8220 0.2735 1.8220 1.3498
No log 0.3214 18 1.5870 0.1636 1.5870 1.2597
No log 0.3571 20 1.5424 0.2456 1.5424 1.2419
No log 0.3929 22 1.8508 0.3846 1.8508 1.3604
No log 0.4286 24 2.4518 0.1744 2.4518 1.5658
No log 0.4643 26 2.0629 0.2857 2.0629 1.4363
No log 0.5 28 1.5241 0.3306 1.5241 1.2345
No log 0.5357 30 1.5456 0.4 1.5456 1.2432
No log 0.5714 32 1.6179 0.4375 1.6179 1.2720
No log 0.6071 34 1.5806 0.4341 1.5806 1.2572
No log 0.6429 36 1.5571 0.4341 1.5571 1.2478
No log 0.6786 38 1.2685 0.4923 1.2685 1.1263
No log 0.7143 40 1.0273 0.528 1.0273 1.0136
No log 0.75 42 1.1676 0.4878 1.1676 1.0805
No log 0.7857 44 1.3294 0.4370 1.3294 1.1530
No log 0.8214 46 1.2542 0.512 1.2542 1.1199
No log 0.8571 48 1.1512 0.5271 1.1512 1.0730
No log 0.8929 50 0.9714 0.6119 0.9714 0.9856
No log 0.9286 52 0.9790 0.6197 0.9790 0.9894
No log 0.9643 54 0.9702 0.6 0.9702 0.9850
No log 1.0 56 0.8595 0.6571 0.8595 0.9271
No log 1.0357 58 0.9409 0.6331 0.9409 0.9700
No log 1.0714 60 0.8638 0.6667 0.8638 0.9294
No log 1.1071 62 1.0278 0.6176 1.0278 1.0138
No log 1.1429 64 1.0168 0.6131 1.0168 1.0083
No log 1.1786 66 0.9303 0.6087 0.9303 0.9645
No log 1.2143 68 1.0525 0.6232 1.0525 1.0259
No log 1.25 70 0.9513 0.6395 0.9513 0.9754
No log 1.2857 72 0.9130 0.6358 0.9130 0.9555
No log 1.3214 74 1.0618 0.6707 1.0618 1.0304
No log 1.3571 76 1.0752 0.6788 1.0752 1.0369
No log 1.3929 78 0.8048 0.7308 0.8048 0.8971
No log 1.4286 80 0.7543 0.6755 0.7543 0.8685
No log 1.4643 82 1.3119 0.5211 1.3119 1.1454
No log 1.5 84 1.4168 0.5211 1.4168 1.1903
No log 1.5357 86 0.8251 0.6923 0.8251 0.9084
No log 1.5714 88 0.6772 0.7791 0.6772 0.8229
No log 1.6071 90 0.7855 0.7374 0.7855 0.8863
No log 1.6429 92 0.5954 0.7702 0.5954 0.7716
No log 1.6786 94 1.0897 0.5286 1.0897 1.0439
No log 1.7143 96 1.3940 0.5070 1.3940 1.1807
No log 1.75 98 1.3117 0.5324 1.3117 1.1453
No log 1.7857 100 1.0506 0.5985 1.0506 1.0250
No log 1.8214 102 1.0008 0.6571 1.0008 1.0004
No log 1.8571 104 0.9999 0.6383 0.9999 1.0000
No log 1.8929 106 0.8517 0.6806 0.8517 0.9229
No log 1.9286 108 0.8864 0.6377 0.8864 0.9415
No log 1.9643 110 0.9238 0.6765 0.9238 0.9611
No log 2.0 112 0.9698 0.6131 0.9698 0.9848
No log 2.0357 114 0.9282 0.6131 0.9282 0.9634
No log 2.0714 116 0.8033 0.6479 0.8033 0.8963
No log 2.1071 118 0.8469 0.6525 0.8469 0.9203
No log 2.1429 120 0.7650 0.7067 0.7650 0.8746
No log 2.1786 122 0.7914 0.7347 0.7914 0.8896
No log 2.2143 124 0.8137 0.7347 0.8137 0.9021
No log 2.25 126 0.8260 0.6713 0.8260 0.9089
No log 2.2857 128 1.2801 0.5828 1.2801 1.1314
No log 2.3214 130 1.2796 0.6 1.2796 1.1312
No log 2.3571 132 0.9335 0.6712 0.9335 0.9662
No log 2.3929 134 0.7317 0.7483 0.7317 0.8554
No log 2.4286 136 0.6986 0.7703 0.6986 0.8358
No log 2.4643 138 0.8018 0.7237 0.8018 0.8954
No log 2.5 140 0.7522 0.7792 0.7522 0.8673
No log 2.5357 142 0.7328 0.7771 0.7328 0.8560
No log 2.5714 144 0.7464 0.7561 0.7464 0.8640
No log 2.6071 146 0.6734 0.7927 0.6734 0.8206
No log 2.6429 148 0.8144 0.6846 0.8144 0.9025
No log 2.6786 150 0.9247 0.6622 0.9247 0.9616
No log 2.7143 152 0.6849 0.7162 0.6849 0.8276
No log 2.75 154 0.6241 0.7724 0.6241 0.7900
No log 2.7857 156 0.6223 0.7945 0.6223 0.7889
No log 2.8214 158 0.6530 0.7843 0.6530 0.8081
No log 2.8571 160 0.7672 0.7013 0.7672 0.8759
No log 2.8929 162 0.8803 0.6962 0.8803 0.9383
No log 2.9286 164 0.6925 0.7843 0.6925 0.8322
No log 2.9643 166 0.6839 0.7619 0.6839 0.8270
No log 3.0 168 0.7331 0.6928 0.7331 0.8562
No log 3.0357 170 0.7334 0.7089 0.7334 0.8564
No log 3.0714 172 0.6745 0.7436 0.6745 0.8213
No log 3.1071 174 0.6469 0.8025 0.6469 0.8043
No log 3.1429 176 0.6549 0.7974 0.6549 0.8093
No log 3.1786 178 0.6765 0.7755 0.6765 0.8225
No log 3.2143 180 0.7089 0.7755 0.7089 0.8420
No log 3.25 182 0.6148 0.8153 0.6148 0.7841
No log 3.2857 184 0.7627 0.7317 0.7627 0.8733
No log 3.3214 186 0.7448 0.7590 0.7448 0.8630
No log 3.3571 188 0.6019 0.8075 0.6019 0.7759
No log 3.3929 190 0.5911 0.8158 0.5911 0.7688
No log 3.4286 192 0.6029 0.8079 0.6029 0.7765
No log 3.4643 194 0.5991 0.8158 0.5991 0.7740
No log 3.5 196 0.5905 0.8158 0.5905 0.7684
No log 3.5357 198 0.6259 0.7919 0.6259 0.7911
No log 3.5714 200 0.7050 0.7448 0.7050 0.8396
No log 3.6071 202 0.7808 0.6993 0.7808 0.8836
No log 3.6429 204 0.8645 0.6622 0.8645 0.9298
No log 3.6786 206 0.7999 0.7013 0.7999 0.8944
No log 3.7143 208 0.6869 0.7383 0.6869 0.8288
No log 3.75 210 0.6098 0.7843 0.6098 0.7809
No log 3.7857 212 0.5783 0.8077 0.5783 0.7604
No log 3.8214 214 0.5961 0.7763 0.5961 0.7721
No log 3.8571 216 0.7018 0.7413 0.7018 0.8377
No log 3.8929 218 0.8312 0.6957 0.8312 0.9117
No log 3.9286 220 0.7869 0.7153 0.7869 0.8871
No log 3.9643 222 0.8012 0.6963 0.8012 0.8951
No log 4.0 224 0.8762 0.6718 0.8762 0.9360
No log 4.0357 226 0.8837 0.6202 0.8837 0.9400
No log 4.0714 228 0.8396 0.6364 0.8396 0.9163
No log 4.1071 230 0.7829 0.6861 0.7829 0.8848
No log 4.1429 232 0.7491 0.7338 0.7491 0.8655
No log 4.1786 234 0.7195 0.7338 0.7195 0.8482
No log 4.2143 236 0.8057 0.6714 0.8057 0.8976
No log 4.25 238 0.7830 0.6950 0.7830 0.8848
No log 4.2857 240 0.6644 0.7483 0.6644 0.8151
No log 4.3214 242 0.6103 0.7919 0.6103 0.7812
No log 4.3571 244 0.6330 0.76 0.6330 0.7956
No log 4.3929 246 0.6843 0.7838 0.6843 0.8272
No log 4.4286 248 0.6849 0.7785 0.6849 0.8276
No log 4.4643 250 0.6814 0.7922 0.6814 0.8255
No log 4.5 252 0.7359 0.7089 0.7359 0.8578
No log 4.5357 254 0.6822 0.7403 0.6822 0.8259
No log 4.5714 256 0.7113 0.7417 0.7113 0.8434
No log 4.6071 258 0.8687 0.6892 0.8687 0.9320
No log 4.6429 260 0.8872 0.6759 0.8872 0.9419
No log 4.6786 262 0.7659 0.7483 0.7659 0.8752
No log 4.7143 264 0.7538 0.7432 0.7538 0.8682
No log 4.75 266 0.8069 0.6309 0.8069 0.8983
No log 4.7857 268 0.7172 0.7467 0.7172 0.8469
No log 4.8214 270 0.7315 0.7534 0.7315 0.8553
No log 4.8571 272 0.8055 0.7123 0.8055 0.8975
No log 4.8929 274 0.7393 0.7552 0.7393 0.8598
No log 4.9286 276 0.7838 0.6928 0.7838 0.8853
No log 4.9643 278 0.8443 0.6490 0.8443 0.9189
No log 5.0 280 0.7595 0.6928 0.7595 0.8715
No log 5.0357 282 0.6939 0.7568 0.6939 0.8330
No log 5.0714 284 0.6802 0.7724 0.6802 0.8248
No log 5.1071 286 0.6864 0.7724 0.6864 0.8285
No log 5.1429 288 0.6460 0.7808 0.6460 0.8037
No log 5.1786 290 0.7898 0.6389 0.7898 0.8887
No log 5.2143 292 0.8585 0.6621 0.8585 0.9266
No log 5.25 294 0.8537 0.6621 0.8537 0.9240
No log 5.2857 296 0.6815 0.7651 0.6815 0.8255
No log 5.3214 298 0.5923 0.8079 0.5923 0.7696
No log 5.3571 300 0.5666 0.8079 0.5666 0.7527
No log 5.3929 302 0.5719 0.7895 0.5719 0.7562
No log 5.4286 304 0.6022 0.7895 0.6022 0.7760
No log 5.4643 306 0.6239 0.7682 0.6239 0.7899
No log 5.5 308 0.6645 0.7746 0.6645 0.8152
No log 5.5357 310 0.7519 0.7246 0.7519 0.8671
No log 5.5714 312 0.8214 0.6667 0.8214 0.9063
No log 5.6071 314 0.8216 0.6462 0.8216 0.9064
No log 5.6429 316 0.8082 0.6269 0.8082 0.8990
No log 5.6786 318 0.7457 0.6667 0.7457 0.8635
No log 5.7143 320 0.6697 0.7778 0.6697 0.8183
No log 5.75 322 0.6919 0.7619 0.6919 0.8318
No log 5.7857 324 0.7928 0.7297 0.7928 0.8904
No log 5.8214 326 0.7842 0.7297 0.7842 0.8855
No log 5.8571 328 0.7699 0.7534 0.7699 0.8774
No log 5.8929 330 0.7298 0.7703 0.7298 0.8543
No log 5.9286 332 0.7266 0.7376 0.7266 0.8524
No log 5.9643 334 0.7869 0.7194 0.7869 0.8871
No log 6.0 336 0.9158 0.7143 0.9158 0.9570
No log 6.0357 338 0.9961 0.6207 0.9961 0.9981
No log 6.0714 340 0.9624 0.625 0.9624 0.9810
No log 6.1071 342 0.8445 0.7143 0.8445 0.9189
No log 6.1429 344 0.7502 0.7143 0.7502 0.8661
No log 6.1786 346 0.7545 0.6957 0.7545 0.8686
No log 6.2143 348 0.7443 0.7101 0.7443 0.8627
No log 6.25 350 0.7063 0.7286 0.7063 0.8404
No log 6.2857 352 0.7408 0.7068 0.7408 0.8607
No log 6.3214 354 0.7287 0.7246 0.7287 0.8536
No log 6.3571 356 0.7529 0.7101 0.7529 0.8677
No log 6.3929 358 0.8052 0.6269 0.8052 0.8973
No log 6.4286 360 0.7469 0.7338 0.7469 0.8642
No log 6.4643 362 0.7481 0.7297 0.7481 0.8649
No log 6.5 364 0.7730 0.7114 0.7730 0.8792
No log 6.5357 366 0.6895 0.7448 0.6895 0.8304
No log 6.5714 368 0.6904 0.7651 0.6904 0.8309
No log 6.6071 370 0.7895 0.7190 0.7895 0.8885
No log 6.6429 372 0.7750 0.6944 0.7750 0.8804
No log 6.6786 374 0.7048 0.7324 0.7048 0.8395
No log 6.7143 376 0.7239 0.7338 0.7239 0.8508
No log 6.75 378 0.8012 0.6861 0.8012 0.8951
No log 6.7857 380 0.7562 0.7260 0.7562 0.8696
No log 6.8214 382 0.6114 0.7448 0.6114 0.7819
No log 6.8571 384 0.6842 0.7516 0.6842 0.8272
No log 6.8929 386 0.8617 0.7044 0.8617 0.9283
No log 6.9286 388 0.8567 0.6667 0.8567 0.9256
No log 6.9643 390 0.7055 0.7297 0.7055 0.8400
No log 7.0 392 0.6905 0.7376 0.6905 0.8310
No log 7.0357 394 0.7482 0.7338 0.7482 0.8650
No log 7.0714 396 0.7352 0.7552 0.7352 0.8574
No log 7.1071 398 0.6614 0.7619 0.6614 0.8132
No log 7.1429 400 0.6964 0.7763 0.6964 0.8345
No log 7.1786 402 0.7235 0.7432 0.7235 0.8506
No log 7.2143 404 0.7783 0.7376 0.7783 0.8822
No log 7.25 406 0.8160 0.7413 0.8160 0.9033
No log 7.2857 408 0.8674 0.6429 0.8674 0.9313
No log 7.3214 410 0.8005 0.7083 0.8005 0.8947
No log 7.3571 412 0.6704 0.7324 0.6704 0.8188
No log 7.3929 414 0.7064 0.6853 0.7064 0.8405
No log 7.4286 416 0.7441 0.6853 0.7441 0.8626
No log 7.4643 418 0.7157 0.6957 0.7157 0.8460
No log 7.5 420 0.7383 0.7111 0.7383 0.8593
No log 7.5357 422 0.7529 0.7111 0.7529 0.8677
No log 7.5714 424 0.7429 0.7050 0.7429 0.8619
No log 7.6071 426 0.7675 0.6222 0.7675 0.8761
No log 7.6429 428 0.7197 0.7586 0.7197 0.8484
No log 7.6786 430 0.6984 0.7413 0.6984 0.8357
No log 7.7143 432 0.7028 0.7286 0.7028 0.8383
No log 7.75 434 0.7192 0.7286 0.7192 0.8480
No log 7.7857 436 0.7019 0.7286 0.7019 0.8378
No log 7.8214 438 0.6752 0.7660 0.6752 0.8217
No log 7.8571 440 0.6547 0.7534 0.6547 0.8091
No log 7.8929 442 0.7021 0.7432 0.7021 0.8379
No log 7.9286 444 0.7029 0.7347 0.7029 0.8384
No log 7.9643 446 0.6542 0.7755 0.6542 0.8088
No log 8.0 448 0.6590 0.75 0.6590 0.8118
No log 8.0357 450 0.7215 0.7172 0.7215 0.8494
No log 8.0714 452 0.7937 0.6950 0.7937 0.8909
No log 8.1071 454 0.7633 0.7153 0.7633 0.8737
No log 8.1429 456 0.7435 0.7246 0.7435 0.8623
No log 8.1786 458 0.7735 0.7164 0.7735 0.8795
No log 8.2143 460 0.8134 0.6667 0.8134 0.9019
No log 8.25 462 0.7708 0.7299 0.7708 0.8780
No log 8.2857 464 0.7072 0.7429 0.7072 0.8410
No log 8.3214 466 0.6698 0.7552 0.6698 0.8184
No log 8.3571 468 0.6989 0.7376 0.6989 0.8360
No log 8.3929 470 0.7752 0.6912 0.7752 0.8804
No log 8.4286 472 0.8575 0.6769 0.8575 0.9260
No log 8.4643 474 0.9081 0.6299 0.9081 0.9530
No log 8.5 476 0.8822 0.6364 0.8822 0.9393
No log 8.5357 478 0.7803 0.7153 0.7803 0.8833
No log 8.5714 480 0.7074 0.7465 0.7074 0.8411
No log 8.6071 482 0.6860 0.7273 0.6860 0.8283
No log 8.6429 484 0.6938 0.7320 0.6938 0.8329
No log 8.6786 486 0.6216 0.76 0.6216 0.7884
No log 8.7143 488 0.6024 0.7973 0.6024 0.7761
No log 8.75 490 0.6194 0.7651 0.6194 0.7870
No log 8.7857 492 0.6416 0.7568 0.6416 0.8010
No log 8.8214 494 0.6116 0.7703 0.6116 0.7821
No log 8.8571 496 0.5887 0.7891 0.5887 0.7673
No log 8.8929 498 0.6002 0.7815 0.6002 0.7747
0.3887 8.9286 500 0.7064 0.7383 0.7064 0.8405
0.3887 8.9643 502 0.7191 0.7226 0.7191 0.8480
0.3887 9.0 504 0.5977 0.7815 0.5977 0.7731
0.3887 9.0357 506 0.5673 0.7919 0.5673 0.7532
0.3887 9.0714 508 0.7052 0.7097 0.7052 0.8398
0.3887 9.1071 510 0.7709 0.6933 0.7709 0.8780
0.3887 9.1429 512 0.6771 0.7237 0.6771 0.8229
0.3887 9.1786 514 0.6017 0.7862 0.6017 0.7757
0.3887 9.2143 516 0.6305 0.7733 0.6305 0.7941
0.3887 9.25 518 0.7318 0.7297 0.7318 0.8555
0.3887 9.2857 520 0.7930 0.6857 0.7930 0.8905
0.3887 9.3214 522 0.8003 0.6618 0.8003 0.8946
0.3887 9.3571 524 0.7984 0.6866 0.7984 0.8935
0.3887 9.3929 526 0.8019 0.6912 0.8019 0.8955

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k12_task1_organization

Finetuned
(4023)
this model