ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k3_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7880
  • Qwk: 0.7007
  • Mse: 0.7880
  • Rmse: 0.8877

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.125 2 7.1782 0.0113 7.1782 2.6792
No log 0.25 4 4.6348 0.0606 4.6348 2.1529
No log 0.375 6 3.0421 0.0988 3.0421 1.7442
No log 0.5 8 2.8462 0.0523 2.8462 1.6871
No log 0.625 10 2.5378 0.0 2.5378 1.5931
No log 0.75 12 2.0376 0.1441 2.0376 1.4274
No log 0.875 14 1.8830 0.1905 1.8830 1.3722
No log 1.0 16 1.8649 0.1538 1.8649 1.3656
No log 1.125 18 1.9883 0.2000 1.9883 1.4101
No log 1.25 20 1.9713 0.2727 1.9713 1.4040
No log 1.375 22 1.7685 0.1538 1.7685 1.3299
No log 1.5 24 1.5656 0.0784 1.5656 1.2512
No log 1.625 26 1.4828 0.1165 1.4828 1.2177
No log 1.75 28 1.5868 0.3509 1.5868 1.2597
No log 1.875 30 1.7250 0.4 1.7250 1.3134
No log 2.0 32 1.7280 0.3937 1.7280 1.3145
No log 2.125 34 1.6467 0.4032 1.6467 1.2832
No log 2.25 36 1.5454 0.4 1.5454 1.2432
No log 2.375 38 1.4445 0.3604 1.4445 1.2019
No log 2.5 40 1.4020 0.2264 1.4020 1.1841
No log 2.625 42 1.4179 0.2617 1.4179 1.1907
No log 2.75 44 1.6025 0.2752 1.6025 1.2659
No log 2.875 46 1.8531 0.3802 1.8531 1.3613
No log 3.0 48 1.8347 0.3443 1.8347 1.3545
No log 3.125 50 1.6767 0.3130 1.6767 1.2949
No log 3.25 52 1.5971 0.3158 1.5971 1.2638
No log 3.375 54 1.5756 0.2679 1.5756 1.2552
No log 3.5 56 1.4754 0.2222 1.4754 1.2147
No log 3.625 58 1.3400 0.2593 1.3400 1.1576
No log 3.75 60 1.2469 0.3243 1.2469 1.1167
No log 3.875 62 1.3548 0.4640 1.3548 1.1639
No log 4.0 64 1.3753 0.4640 1.3753 1.1727
No log 4.125 66 1.2046 0.4677 1.2046 1.0975
No log 4.25 68 1.1218 0.4370 1.1218 1.0592
No log 4.375 70 1.1306 0.3898 1.1306 1.0633
No log 4.5 72 1.0821 0.4874 1.0821 1.0402
No log 4.625 74 1.1538 0.6047 1.1538 1.0741
No log 4.75 76 1.2484 0.4921 1.2484 1.1173
No log 4.875 78 1.1890 0.5038 1.1890 1.0904
No log 5.0 80 1.1219 0.6107 1.1219 1.0592
No log 5.125 82 0.9881 0.6154 0.9881 0.9940
No log 5.25 84 0.9095 0.6154 0.9095 0.9537
No log 5.375 86 0.9036 0.6714 0.9036 0.9506
No log 5.5 88 0.9079 0.6418 0.9079 0.9528
No log 5.625 90 0.9628 0.5891 0.9628 0.9812
No log 5.75 92 1.1488 0.5938 1.1488 1.0718
No log 5.875 94 1.2421 0.5522 1.2421 1.1145
No log 6.0 96 1.2029 0.5735 1.2029 1.0968
No log 6.125 98 1.1919 0.5954 1.1919 1.0917
No log 6.25 100 0.9717 0.5954 0.9717 0.9857
No log 6.375 102 0.7580 0.7286 0.7580 0.8706
No log 6.5 104 0.7103 0.7183 0.7103 0.8428
No log 6.625 106 0.7302 0.7571 0.7302 0.8545
No log 6.75 108 0.8096 0.6815 0.8096 0.8998
No log 6.875 110 0.8827 0.6324 0.8827 0.9395
No log 7.0 112 0.7803 0.6950 0.7803 0.8833
No log 7.125 114 0.7106 0.7042 0.7106 0.8430
No log 7.25 116 0.6987 0.7552 0.6987 0.8359
No log 7.375 118 0.9232 0.6119 0.9232 0.9608
No log 7.5 120 1.2048 0.5038 1.2048 1.0977
No log 7.625 122 1.2263 0.5156 1.2263 1.1074
No log 7.75 124 1.0953 0.5891 1.0953 1.0466
No log 7.875 126 0.9632 0.6061 0.9632 0.9815
No log 8.0 128 0.8625 0.6715 0.8625 0.9287
No log 8.125 130 0.8516 0.7222 0.8516 0.9228
No log 8.25 132 0.8042 0.7222 0.8042 0.8968
No log 8.375 134 0.7791 0.7397 0.7791 0.8827
No log 8.5 136 0.7413 0.7324 0.7413 0.8610
No log 8.625 138 0.7592 0.7092 0.7592 0.8713
No log 8.75 140 0.7869 0.6993 0.7869 0.8871
No log 8.875 142 0.7555 0.7042 0.7555 0.8692
No log 9.0 144 0.6947 0.7133 0.6947 0.8335
No log 9.125 146 0.6686 0.7703 0.6686 0.8177
No log 9.25 148 0.7585 0.7034 0.7585 0.8709
No log 9.375 150 0.7995 0.7034 0.7995 0.8942
No log 9.5 152 0.8118 0.6853 0.8118 0.9010
No log 9.625 154 0.8821 0.6621 0.8821 0.9392
No log 9.75 156 1.0523 0.6259 1.0523 1.0258
No log 9.875 158 1.0136 0.6131 1.0136 1.0068
No log 10.0 160 0.8649 0.6412 0.8649 0.9300
No log 10.125 162 0.8034 0.6667 0.8034 0.8963
No log 10.25 164 0.7518 0.7310 0.7518 0.8670
No log 10.375 166 0.7354 0.7310 0.7354 0.8575
No log 10.5 168 0.7447 0.7310 0.7447 0.8630
No log 10.625 170 0.7578 0.6803 0.7578 0.8705
No log 10.75 172 0.7701 0.7368 0.7701 0.8775
No log 10.875 174 0.6774 0.7432 0.6774 0.8230
No log 11.0 176 0.6078 0.7785 0.6078 0.7796
No log 11.125 178 0.6346 0.7724 0.6346 0.7966
No log 11.25 180 0.7423 0.7286 0.7423 0.8615
No log 11.375 182 0.9601 0.6331 0.9601 0.9798
No log 11.5 184 1.0685 0.5926 1.0685 1.0337
No log 11.625 186 0.9877 0.5606 0.9877 0.9938
No log 11.75 188 0.9523 0.5970 0.9523 0.9758
No log 11.875 190 0.9026 0.6074 0.9026 0.9500
No log 12.0 192 0.8673 0.6812 0.8673 0.9313
No log 12.125 194 0.9628 0.6759 0.9628 0.9812
No log 12.25 196 0.9684 0.6434 0.9684 0.9841
No log 12.375 198 0.9034 0.6143 0.9034 0.9505
No log 12.5 200 0.7984 0.7034 0.7984 0.8936
No log 12.625 202 0.7372 0.7383 0.7372 0.8586
No log 12.75 204 0.7196 0.7285 0.7196 0.8483
No log 12.875 206 0.6691 0.7619 0.6691 0.8180
No log 13.0 208 0.7955 0.75 0.7955 0.8919
No log 13.125 210 0.8343 0.7027 0.8343 0.9134
No log 13.25 212 0.7176 0.7286 0.7176 0.8471
No log 13.375 214 0.6599 0.7808 0.6599 0.8123
No log 13.5 216 0.6615 0.7838 0.6615 0.8134
No log 13.625 218 0.7045 0.7692 0.7045 0.8393
No log 13.75 220 0.7740 0.6950 0.7740 0.8798
No log 13.875 222 0.7850 0.6812 0.7850 0.8860
No log 14.0 224 0.7226 0.7183 0.7226 0.8501
No log 14.125 226 0.6812 0.6713 0.6812 0.8254
No log 14.25 228 0.7211 0.7152 0.7211 0.8492
No log 14.375 230 0.6998 0.7172 0.6998 0.8366
No log 14.5 232 0.7615 0.6957 0.7615 0.8727
No log 14.625 234 0.9316 0.6620 0.9316 0.9652
No log 14.75 236 0.9765 0.6174 0.9765 0.9882
No log 14.875 238 0.8494 0.6475 0.8494 0.9216
No log 15.0 240 0.7488 0.6912 0.7488 0.8653
No log 15.125 242 0.7333 0.7101 0.7333 0.8563
No log 15.25 244 0.7463 0.7007 0.7463 0.8639
No log 15.375 246 0.7833 0.6667 0.7833 0.8851
No log 15.5 248 0.7598 0.6912 0.7598 0.8717
No log 15.625 250 0.7114 0.7050 0.7114 0.8435
No log 15.75 252 0.6872 0.7552 0.6872 0.8290
No log 15.875 254 0.6893 0.7465 0.6893 0.8302
No log 16.0 256 0.7124 0.7286 0.7124 0.8440
No log 16.125 258 0.7309 0.7273 0.7309 0.8549
No log 16.25 260 0.7319 0.7483 0.7319 0.8555
No log 16.375 262 0.7971 0.6832 0.7971 0.8928
No log 16.5 264 0.8941 0.6667 0.8941 0.9456
No log 16.625 266 1.0582 0.6135 1.0582 1.0287
No log 16.75 268 1.0886 0.5987 1.0886 1.0434
No log 16.875 270 1.0127 0.5890 1.0127 1.0063
No log 17.0 272 0.8809 0.6715 0.8809 0.9386
No log 17.125 274 0.8292 0.6901 0.8292 0.9106
No log 17.25 276 0.8386 0.7368 0.8386 0.9158
No log 17.375 278 0.7719 0.76 0.7719 0.8786
No log 17.5 280 0.6983 0.7815 0.6983 0.8356
No log 17.625 282 0.7288 0.76 0.7288 0.8537
No log 17.75 284 0.6835 0.7483 0.6835 0.8268
No log 17.875 286 0.6311 0.7338 0.6311 0.7944
No log 18.0 288 0.6734 0.7101 0.6734 0.8206
No log 18.125 290 0.8069 0.7050 0.8069 0.8983
No log 18.25 292 0.9982 0.6667 0.9982 0.9991
No log 18.375 294 1.0230 0.6624 1.0230 1.0114
No log 18.5 296 0.9284 0.6621 0.9284 0.9636
No log 18.625 298 0.8317 0.7211 0.8317 0.9120
No log 18.75 300 0.8021 0.7133 0.8021 0.8956
No log 18.875 302 0.8352 0.7260 0.8352 0.9139
No log 19.0 304 0.9263 0.6713 0.9263 0.9625
No log 19.125 306 0.9219 0.6531 0.9219 0.9601
No log 19.25 308 0.8928 0.6713 0.8928 0.9449
No log 19.375 310 0.7789 0.7286 0.7789 0.8825
No log 19.5 312 0.7235 0.7465 0.7235 0.8506
No log 19.625 314 0.7313 0.7376 0.7313 0.8551
No log 19.75 316 0.7878 0.7042 0.7878 0.8876
No log 19.875 318 0.7693 0.7361 0.7693 0.8771
No log 20.0 320 0.8107 0.7172 0.8107 0.9004
No log 20.125 322 0.8081 0.7042 0.8081 0.8989
No log 20.25 324 0.7870 0.7034 0.7870 0.8872
No log 20.375 326 0.8578 0.6792 0.8578 0.9262
No log 20.5 328 0.8648 0.7143 0.8648 0.9300
No log 20.625 330 0.7938 0.7034 0.7938 0.8909
No log 20.75 332 0.7466 0.7143 0.7466 0.8640
No log 20.875 334 0.7601 0.6906 0.7601 0.8718
No log 21.0 336 0.8031 0.6957 0.8031 0.8962
No log 21.125 338 0.8329 0.6957 0.8329 0.9126
No log 21.25 340 0.8588 0.6957 0.8588 0.9267
No log 21.375 342 0.8881 0.6471 0.8881 0.9424
No log 21.5 344 0.8469 0.6906 0.8469 0.9203
No log 21.625 346 0.8122 0.7172 0.8122 0.9012
No log 21.75 348 0.7828 0.7133 0.7828 0.8848
No log 21.875 350 0.7495 0.7234 0.7495 0.8657
No log 22.0 352 0.7220 0.7183 0.7220 0.8497
No log 22.125 354 0.7387 0.7050 0.7387 0.8595
No log 22.25 356 0.7980 0.6906 0.7980 0.8933
No log 22.375 358 0.9282 0.7075 0.9282 0.9634
No log 22.5 360 0.9424 0.7075 0.9424 0.9707
No log 22.625 362 0.8319 0.7211 0.8319 0.9121
No log 22.75 364 0.6971 0.7338 0.6971 0.8349
No log 22.875 366 0.6544 0.7376 0.6544 0.8090
No log 23.0 368 0.6450 0.7606 0.6450 0.8031
No log 23.125 370 0.6413 0.7660 0.6413 0.8008
No log 23.25 372 0.6520 0.7571 0.6520 0.8075
No log 23.375 374 0.7075 0.7324 0.7075 0.8411
No log 23.5 376 0.7208 0.7324 0.7208 0.8490
No log 23.625 378 0.6926 0.7101 0.6926 0.8322
No log 23.75 380 0.7095 0.7194 0.7095 0.8423
No log 23.875 382 0.7054 0.7429 0.7054 0.8399
No log 24.0 384 0.7165 0.7101 0.7165 0.8464
No log 24.125 386 0.7877 0.6901 0.7877 0.8876
No log 24.25 388 0.8149 0.7205 0.8149 0.9027
No log 24.375 390 0.7593 0.7089 0.7593 0.8714
No log 24.5 392 0.7027 0.7248 0.7027 0.8383
No log 24.625 394 0.6978 0.7432 0.6978 0.8354
No log 24.75 396 0.7391 0.7 0.7391 0.8597
No log 24.875 398 0.8049 0.6901 0.8049 0.8972
No log 25.0 400 0.8036 0.6713 0.8036 0.8964
No log 25.125 402 0.8000 0.6667 0.8000 0.8944
No log 25.25 404 0.7523 0.7101 0.7523 0.8674
No log 25.375 406 0.7228 0.7101 0.7228 0.8502
No log 25.5 408 0.6994 0.7183 0.6994 0.8363
No log 25.625 410 0.7286 0.7092 0.7286 0.8536
No log 25.75 412 0.7665 0.7361 0.7665 0.8755
No log 25.875 414 0.7953 0.7133 0.7953 0.8918
No log 26.0 416 0.7777 0.7260 0.7777 0.8819
No log 26.125 418 0.7310 0.7234 0.7310 0.8550
No log 26.25 420 0.7383 0.7143 0.7383 0.8593
No log 26.375 422 0.7610 0.7376 0.7610 0.8724
No log 26.5 424 0.8391 0.6667 0.8391 0.9160
No log 26.625 426 0.9234 0.6383 0.9234 0.9609
No log 26.75 428 0.9510 0.6 0.9510 0.9752
No log 26.875 430 0.9421 0.6377 0.9421 0.9706
No log 27.0 432 0.8800 0.6567 0.8800 0.9381
No log 27.125 434 0.7857 0.6667 0.7857 0.8864
No log 27.25 436 0.7497 0.7153 0.7497 0.8659
No log 27.375 438 0.7003 0.7338 0.7003 0.8368
No log 27.5 440 0.6944 0.7465 0.6944 0.8333
No log 27.625 442 0.7350 0.7448 0.7350 0.8573
No log 27.75 444 0.7287 0.7448 0.7287 0.8536
No log 27.875 446 0.6968 0.7338 0.6968 0.8348
No log 28.0 448 0.6957 0.7660 0.6957 0.8341
No log 28.125 450 0.7286 0.7246 0.7286 0.8536
No log 28.25 452 0.8045 0.6901 0.8045 0.8969
No log 28.375 454 0.8098 0.7013 0.8098 0.8999
No log 28.5 456 0.7280 0.7347 0.7280 0.8532
No log 28.625 458 0.6901 0.7639 0.6901 0.8307
No log 28.75 460 0.7152 0.7234 0.7152 0.8457
No log 28.875 462 0.7837 0.6906 0.7837 0.8853
No log 29.0 464 0.7788 0.7153 0.7788 0.8825
No log 29.125 466 0.7467 0.7338 0.7467 0.8641
No log 29.25 468 0.7069 0.7571 0.7069 0.8408
No log 29.375 470 0.7221 0.7571 0.7221 0.8497
No log 29.5 472 0.7198 0.7660 0.7198 0.8484
No log 29.625 474 0.7086 0.7639 0.7086 0.8418
No log 29.75 476 0.7414 0.75 0.7414 0.8611
No log 29.875 478 0.8881 0.6667 0.8881 0.9424
No log 30.0 480 0.9443 0.6587 0.9443 0.9718
No log 30.125 482 0.8618 0.6829 0.8618 0.9283
No log 30.25 484 0.7697 0.7417 0.7697 0.8773
No log 30.375 486 0.7375 0.7324 0.7375 0.8588
No log 30.5 488 0.7565 0.7246 0.7565 0.8698
No log 30.625 490 0.7854 0.7153 0.7854 0.8862
No log 30.75 492 0.8088 0.6815 0.8088 0.8993
No log 30.875 494 0.8165 0.6815 0.8165 0.9036
No log 31.0 496 0.7959 0.7246 0.7959 0.8921
No log 31.125 498 0.7368 0.7413 0.7368 0.8584
0.393 31.25 500 0.6891 0.7413 0.6891 0.8301
0.393 31.375 502 0.6864 0.7413 0.6864 0.8285
0.393 31.5 504 0.7375 0.7619 0.7375 0.8588
0.393 31.625 506 0.8386 0.6795 0.8386 0.9157
0.393 31.75 508 0.8663 0.6497 0.8663 0.9308
0.393 31.875 510 0.8101 0.6993 0.8101 0.9000
0.393 32.0 512 0.7784 0.7 0.7784 0.8823
0.393 32.125 514 0.7880 0.7007 0.7880 0.8877

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run3_AugV5_k3_task1_organization

Finetuned
(4023)
this model