ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k8_task1_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8957
  • Qwk: 0.6522
  • Mse: 0.8957
  • Rmse: 0.9464

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0526 2 7.3686 0.0 7.3686 2.7145
No log 0.1053 4 4.4678 0.0412 4.4678 2.1137
No log 0.1579 6 3.1469 0.0889 3.1469 1.7739
No log 0.2105 8 2.3633 0.1477 2.3633 1.5373
No log 0.2632 10 1.8118 0.1947 1.8118 1.3460
No log 0.3158 12 1.7299 0.2478 1.7299 1.3153
No log 0.3684 14 1.8190 0.1897 1.8190 1.3487
No log 0.4211 16 1.8961 0.2167 1.8961 1.3770
No log 0.4737 18 1.8248 0.3025 1.8248 1.3509
No log 0.5263 20 1.9064 0.2479 1.9064 1.3807
No log 0.5789 22 1.7364 0.2807 1.7364 1.3177
No log 0.6316 24 1.6775 0.2385 1.6775 1.2952
No log 0.6842 26 1.5300 0.2243 1.5300 1.2369
No log 0.7368 28 1.5036 0.1524 1.5036 1.2262
No log 0.7895 30 1.5175 0.2243 1.5175 1.2319
No log 0.8421 32 1.5450 0.2075 1.5450 1.2430
No log 0.8947 34 1.5743 0.2056 1.5743 1.2547
No log 0.9474 36 1.6203 0.2414 1.6203 1.2729
No log 1.0 38 1.4762 0.2162 1.4762 1.2150
No log 1.0526 40 1.4251 0.2018 1.4251 1.1938
No log 1.1053 42 1.4139 0.2301 1.4139 1.1891
No log 1.1579 44 1.4442 0.2735 1.4442 1.2017
No log 1.2105 46 1.3150 0.3761 1.3150 1.1467
No log 1.2632 48 1.3332 0.2407 1.3332 1.1546
No log 1.3158 50 1.3115 0.3363 1.3115 1.1452
No log 1.3684 52 1.3765 0.3471 1.3765 1.1732
No log 1.4211 54 1.3550 0.3680 1.3550 1.1641
No log 1.4737 56 1.2741 0.4160 1.2741 1.1287
No log 1.5263 58 1.0920 0.4878 1.0920 1.0450
No log 1.5789 60 1.0072 0.6047 1.0072 1.0036
No log 1.6316 62 1.0038 0.5600 1.0038 1.0019
No log 1.6842 64 1.0134 0.5984 1.0134 1.0067
No log 1.7368 66 1.0525 0.528 1.0525 1.0259
No log 1.7895 68 0.9353 0.6154 0.9353 0.9671
No log 1.8421 70 0.8940 0.6364 0.8940 0.9455
No log 1.8947 72 1.3071 0.4812 1.3071 1.1433
No log 1.9474 74 1.4288 0.4409 1.4288 1.1953
No log 2.0 76 1.2336 0.5077 1.2336 1.1107
No log 2.0526 78 1.0872 0.6567 1.0872 1.0427
No log 2.1053 80 0.9393 0.6617 0.9393 0.9692
No log 2.1579 82 0.9283 0.6515 0.9283 0.9635
No log 2.2105 84 1.0614 0.6094 1.0614 1.0302
No log 2.2632 86 1.2628 0.6094 1.2628 1.1238
No log 2.3158 88 1.2678 0.5303 1.2678 1.1259
No log 2.3684 90 1.1243 0.5630 1.1243 1.0604
No log 2.4211 92 0.9874 0.6519 0.9874 0.9937
No log 2.4737 94 0.9815 0.6866 0.9815 0.9907
No log 2.5263 96 1.1327 0.5806 1.1327 1.0643
No log 2.5789 98 1.2102 0.5410 1.2102 1.1001
No log 2.6316 100 1.1408 0.6615 1.1408 1.0681
No log 2.6842 102 1.1131 0.5778 1.1131 1.0551
No log 2.7368 104 1.2031 0.5373 1.2031 1.0969
No log 2.7895 106 1.2385 0.5147 1.2385 1.1129
No log 2.8421 108 1.0806 0.5612 1.0806 1.0395
No log 2.8947 110 0.9418 0.6519 0.9418 0.9705
No log 2.9474 112 0.8949 0.6107 0.8949 0.9460
No log 3.0 114 0.8719 0.6565 0.8719 0.9338
No log 3.0526 116 0.9020 0.6462 0.9020 0.9498
No log 3.1053 118 1.0699 0.6349 1.0699 1.0344
No log 3.1579 120 1.0895 0.6 1.0895 1.0438
No log 3.2105 122 0.9558 0.6515 0.9558 0.9777
No log 3.2632 124 0.7873 0.7206 0.7873 0.8873
No log 3.3158 126 0.7802 0.7015 0.7802 0.8833
No log 3.3684 128 0.8364 0.6667 0.8364 0.9146
No log 3.4211 130 0.9262 0.6818 0.9262 0.9624
No log 3.4737 132 1.0389 0.6316 1.0389 1.0193
No log 3.5263 134 1.0176 0.6316 1.0176 1.0088
No log 3.5789 136 1.0638 0.6032 1.0638 1.0314
No log 3.6316 138 1.1493 0.4310 1.1493 1.0720
No log 3.6842 140 1.2320 0.3540 1.2320 1.1100
No log 3.7368 142 1.1178 0.3604 1.1178 1.0572
No log 3.7895 144 0.9088 0.6142 0.9088 0.9533
No log 3.8421 146 0.7565 0.7571 0.7565 0.8698
No log 3.8947 148 0.7734 0.7059 0.7734 0.8794
No log 3.9474 150 0.7034 0.7832 0.7034 0.8387
No log 4.0 152 0.7390 0.7552 0.7390 0.8596
No log 4.0526 154 0.8745 0.6519 0.8745 0.9351
No log 4.1053 156 1.0781 0.5846 1.0781 1.0383
No log 4.1579 158 1.1305 0.5625 1.1305 1.0632
No log 4.2105 160 1.0967 0.5410 1.0967 1.0472
No log 4.2632 162 0.9234 0.6423 0.9234 0.9609
No log 4.3158 164 0.8359 0.6901 0.8359 0.9143
No log 4.3684 166 0.8105 0.7042 0.8105 0.9003
No log 4.4211 168 0.8678 0.6812 0.8678 0.9316
No log 4.4737 170 0.9283 0.6619 0.9283 0.9635
No log 4.5263 172 0.8601 0.6761 0.8601 0.9274
No log 4.5789 174 0.8761 0.6809 0.8761 0.9360
No log 4.6316 176 0.8701 0.6713 0.8701 0.9328
No log 4.6842 178 0.8336 0.6944 0.8336 0.9130
No log 4.7368 180 0.8581 0.6522 0.8581 0.9263
No log 4.7895 182 0.9203 0.6202 0.9203 0.9593
No log 4.8421 184 0.9882 0.5366 0.9882 0.9941
No log 4.8947 186 0.9923 0.6094 0.9923 0.9961
No log 4.9474 188 0.9447 0.6074 0.9447 0.9719
No log 5.0 190 0.8256 0.6667 0.8256 0.9086
No log 5.0526 192 0.7592 0.7075 0.7592 0.8713
No log 5.1053 194 0.7687 0.7075 0.7687 0.8767
No log 5.1579 196 0.7829 0.6800 0.7829 0.8848
No log 5.2105 198 0.8941 0.6165 0.8941 0.9456
No log 5.2632 200 0.9799 0.6119 0.9799 0.9899
No log 5.3158 202 0.9258 0.6260 0.9258 0.9622
No log 5.3684 204 0.8217 0.6299 0.8217 0.9065
No log 5.4211 206 0.7824 0.6176 0.7824 0.8846
No log 5.4737 208 0.7446 0.7 0.7446 0.8629
No log 5.5263 210 0.7696 0.6714 0.7696 0.8773
No log 5.5789 212 0.8408 0.6471 0.8408 0.9169
No log 5.6316 214 0.9245 0.6370 0.9245 0.9615
No log 5.6842 216 0.9009 0.6522 0.9009 0.9491
No log 5.7368 218 0.8637 0.6619 0.8637 0.9293
No log 5.7895 220 0.8565 0.6619 0.8565 0.9255
No log 5.8421 222 0.9315 0.6331 0.9315 0.9651
No log 5.8947 224 1.0412 0.6197 1.0412 1.0204
No log 5.9474 226 1.0841 0.5816 1.0841 1.0412
No log 6.0 228 1.0784 0.5833 1.0784 1.0385
No log 6.0526 230 1.0379 0.6111 1.0379 1.0188
No log 6.1053 232 0.8824 0.6528 0.8824 0.9394
No log 6.1579 234 0.7992 0.7297 0.7992 0.8940
No log 6.2105 236 0.8010 0.7297 0.8010 0.8950
No log 6.2632 238 0.8521 0.6857 0.8521 0.9231
No log 6.3158 240 0.8656 0.6618 0.8656 0.9304
No log 6.3684 242 0.8253 0.7083 0.8253 0.9085
No log 6.4211 244 0.7672 0.7432 0.7672 0.8759
No log 6.4737 246 0.7567 0.7162 0.7567 0.8699
No log 6.5263 248 0.7671 0.7162 0.7671 0.8758
No log 6.5789 250 0.7849 0.6993 0.7849 0.8860
No log 6.6316 252 0.7979 0.6901 0.7979 0.8933
No log 6.6842 254 0.8366 0.6901 0.8366 0.9147
No log 6.7368 256 0.8519 0.6531 0.8519 0.9230
No log 6.7895 258 0.7782 0.6846 0.7782 0.8822
No log 6.8421 260 0.7346 0.7662 0.7346 0.8571
No log 6.8947 262 0.7298 0.7467 0.7298 0.8543
No log 6.9474 264 0.7357 0.7432 0.7357 0.8577
No log 7.0 266 0.7246 0.7248 0.7246 0.8512
No log 7.0526 268 0.6738 0.7771 0.6738 0.8208
No log 7.1053 270 0.6437 0.8176 0.6437 0.8023
No log 7.1579 272 0.6453 0.8025 0.6453 0.8033
No log 7.2105 274 0.6523 0.8 0.6523 0.8077
No log 7.2632 276 0.6789 0.7733 0.6789 0.8239
No log 7.3158 278 0.7206 0.7273 0.7206 0.8489
No log 7.3684 280 0.7907 0.6812 0.7907 0.8892
No log 7.4211 282 0.7855 0.6803 0.7855 0.8863
No log 7.4737 284 0.7385 0.6797 0.7385 0.8593
No log 7.5263 286 0.7507 0.6715 0.7507 0.8664
No log 7.5789 288 0.7722 0.6718 0.7722 0.8788
No log 7.6316 290 0.7972 0.6615 0.7972 0.8929
No log 7.6842 292 0.7910 0.6718 0.7910 0.8894
No log 7.7368 294 0.8079 0.6667 0.8079 0.8988
No log 7.7895 296 0.8392 0.6331 0.8392 0.9161
No log 7.8421 298 0.8185 0.6429 0.8185 0.9047
No log 7.8947 300 0.8142 0.6423 0.8142 0.9023
No log 7.9474 302 0.8252 0.6715 0.8252 0.9084
No log 8.0 304 0.8643 0.6963 0.8643 0.9297
No log 8.0526 306 0.8691 0.7015 0.8691 0.9322
No log 8.1053 308 0.8314 0.6870 0.8314 0.9118
No log 8.1579 310 0.7646 0.6963 0.7646 0.8744
No log 8.2105 312 0.7319 0.6861 0.7319 0.8555
No log 8.2632 314 0.7196 0.7172 0.7196 0.8483
No log 8.3158 316 0.6796 0.7483 0.6796 0.8244
No log 8.3684 318 0.7075 0.7260 0.7075 0.8412
No log 8.4211 320 0.7639 0.72 0.7639 0.8740
No log 8.4737 322 0.7449 0.7133 0.7449 0.8631
No log 8.5263 324 0.7303 0.7059 0.7303 0.8546
No log 8.5789 326 0.7523 0.7222 0.7523 0.8674
No log 8.6316 328 0.7518 0.7042 0.7518 0.8671
No log 8.6842 330 0.7256 0.7310 0.7256 0.8518
No log 8.7368 332 0.6936 0.7712 0.6936 0.8328
No log 8.7895 334 0.7035 0.7771 0.7035 0.8388
No log 8.8421 336 0.7095 0.7643 0.7095 0.8423
No log 8.8947 338 0.8056 0.7152 0.8056 0.8975
No log 8.9474 340 0.8375 0.7013 0.8375 0.9151
No log 9.0 342 0.7598 0.7324 0.7598 0.8717
No log 9.0526 344 0.7429 0.7133 0.7429 0.8619
No log 9.1053 346 0.7046 0.7432 0.7046 0.8394
No log 9.1579 348 0.6619 0.7432 0.6619 0.8136
No log 9.2105 350 0.6755 0.7703 0.6755 0.8219
No log 9.2632 352 0.7373 0.7619 0.7373 0.8586
No log 9.3158 354 0.7960 0.7143 0.7960 0.8922
No log 9.3684 356 0.7951 0.7190 0.7951 0.8917
No log 9.4211 358 0.7157 0.7324 0.7157 0.8460
No log 9.4737 360 0.6921 0.7133 0.6921 0.8319
No log 9.5263 362 0.6950 0.7347 0.6950 0.8337
No log 9.5789 364 0.6915 0.7042 0.6915 0.8316
No log 9.6316 366 0.7165 0.7007 0.7165 0.8465
No log 9.6842 368 0.7219 0.7246 0.7219 0.8497
No log 9.7368 370 0.7275 0.6957 0.7275 0.8529
No log 9.7895 372 0.7442 0.6957 0.7442 0.8627
No log 9.8421 374 0.7792 0.6912 0.7792 0.8827
No log 9.8947 376 0.8423 0.6917 0.8423 0.9178
No log 9.9474 378 0.8592 0.6818 0.8592 0.9269
No log 10.0 380 0.7795 0.6866 0.7795 0.8829
No log 10.0526 382 0.7406 0.6906 0.7406 0.8606
No log 10.1053 384 0.7538 0.6957 0.7538 0.8682
No log 10.1579 386 0.7372 0.7194 0.7372 0.8586
No log 10.2105 388 0.7394 0.6957 0.7394 0.8599
No log 10.2632 390 0.7533 0.6957 0.7533 0.8679
No log 10.3158 392 0.7540 0.6957 0.7540 0.8684
No log 10.3684 394 0.7483 0.6957 0.7483 0.8651
No log 10.4211 396 0.7367 0.6906 0.7367 0.8583
No log 10.4737 398 0.7632 0.7092 0.7632 0.8736
No log 10.5263 400 0.7552 0.7310 0.7552 0.8690
No log 10.5789 402 0.7346 0.7651 0.7346 0.8571
No log 10.6316 404 0.7622 0.7383 0.7622 0.8730
No log 10.6842 406 0.8533 0.6479 0.8533 0.9237
No log 10.7368 408 0.8187 0.6950 0.8187 0.9048
No log 10.7895 410 0.7702 0.6853 0.7702 0.8776
No log 10.8421 412 0.7376 0.6986 0.7376 0.8588
No log 10.8947 414 0.6694 0.7532 0.6694 0.8182
No log 10.9474 416 0.6421 0.7871 0.6421 0.8013
No log 11.0 418 0.6562 0.7771 0.6562 0.8100
No log 11.0526 420 0.6909 0.7771 0.6909 0.8312
No log 11.1053 422 0.7465 0.7134 0.7465 0.8640
No log 11.1579 424 0.7735 0.6933 0.7735 0.8795
No log 11.2105 426 0.8270 0.6901 0.8270 0.9094
No log 11.2632 428 0.8049 0.6370 0.8049 0.8971
No log 11.3158 430 0.7463 0.7164 0.7463 0.8639
No log 11.3684 432 0.7037 0.7007 0.7037 0.8389
No log 11.4211 434 0.6453 0.7273 0.6453 0.8033
No log 11.4737 436 0.5885 0.7632 0.5885 0.7671
No log 11.5263 438 0.5565 0.7821 0.5565 0.7460
No log 11.5789 440 0.5589 0.7975 0.5589 0.7476
No log 11.6316 442 0.6105 0.7516 0.6105 0.7813
No log 11.6842 444 0.6591 0.7075 0.6591 0.8118
No log 11.7368 446 0.6948 0.6803 0.6948 0.8336
No log 11.7895 448 0.6895 0.6901 0.6895 0.8304
No log 11.8421 450 0.6700 0.7376 0.6700 0.8185
No log 11.8947 452 0.6858 0.7222 0.6858 0.8281
No log 11.9474 454 0.7841 0.7089 0.7841 0.8855
No log 12.0 456 0.8076 0.7089 0.8076 0.8987
No log 12.0526 458 0.7724 0.6901 0.7724 0.8789
No log 12.1053 460 0.7815 0.6618 0.7815 0.8840
No log 12.1579 462 0.7478 0.6667 0.7478 0.8647
No log 12.2105 464 0.7336 0.7042 0.7336 0.8565
No log 12.2632 466 0.7572 0.6667 0.7572 0.8702
No log 12.3158 468 0.8159 0.6519 0.8159 0.9033
No log 12.3684 470 0.8356 0.6418 0.8356 0.9141
No log 12.4211 472 0.8327 0.6519 0.8327 0.9125
No log 12.4737 474 0.8076 0.6763 0.8076 0.8987
No log 12.5263 476 0.8034 0.7273 0.8034 0.8963
No log 12.5789 478 0.8481 0.6795 0.8481 0.9209
No log 12.6316 480 0.8165 0.6974 0.8165 0.9036
No log 12.6842 482 0.7549 0.6761 0.7549 0.8689
No log 12.7368 484 0.7487 0.7143 0.7487 0.8652
No log 12.7895 486 0.7611 0.7007 0.7611 0.8724
No log 12.8421 488 0.7927 0.6957 0.7927 0.8903
No log 12.8947 490 0.8211 0.6667 0.8211 0.9062
No log 12.9474 492 0.8918 0.6389 0.8918 0.9444
No log 13.0 494 0.9870 0.6587 0.9870 0.9935
No log 13.0526 496 0.9341 0.6667 0.9341 0.9665
No log 13.1053 498 0.8365 0.6667 0.8365 0.9146
0.4162 13.1579 500 0.7641 0.6939 0.7641 0.8741
0.4162 13.2105 502 0.7313 0.7075 0.7313 0.8552
0.4162 13.2632 504 0.7189 0.6939 0.7189 0.8479
0.4162 13.3158 506 0.7758 0.6839 0.7758 0.8808
0.4162 13.3684 508 0.7692 0.6839 0.7692 0.8770
0.4162 13.4211 510 0.7140 0.7059 0.7140 0.8450
0.4162 13.4737 512 0.6602 0.7368 0.6602 0.8125
0.4162 13.5263 514 0.6683 0.7755 0.6683 0.8175
0.4162 13.5789 516 0.6763 0.7639 0.6763 0.8224
0.4162 13.6316 518 0.6613 0.7639 0.6613 0.8132
0.4162 13.6842 520 0.6530 0.7297 0.6530 0.8081
0.4162 13.7368 522 0.6998 0.7105 0.6998 0.8365
0.4162 13.7895 524 0.6946 0.7097 0.6946 0.8334
0.4162 13.8421 526 0.6726 0.7329 0.6726 0.8201
0.4162 13.8947 528 0.6716 0.7531 0.6716 0.8195
0.4162 13.9474 530 0.6787 0.7417 0.6787 0.8238
0.4162 14.0 532 0.7358 0.6993 0.7358 0.8578
0.4162 14.0526 534 0.8526 0.6531 0.8526 0.9233
0.4162 14.1053 536 0.8820 0.6533 0.8820 0.9392
0.4162 14.1579 538 0.8174 0.6842 0.8174 0.9041
0.4162 14.2105 540 0.7616 0.7248 0.7616 0.8727
0.4162 14.2632 542 0.7056 0.7248 0.7056 0.8400
0.4162 14.3158 544 0.7032 0.7248 0.7032 0.8385
0.4162 14.3684 546 0.7293 0.7237 0.7293 0.8540
0.4162 14.4211 548 0.7903 0.725 0.7903 0.8890
0.4162 14.4737 550 0.8414 0.7170 0.8414 0.9173
0.4162 14.5263 552 0.9320 0.6871 0.9320 0.9654
0.4162 14.5789 554 0.9700 0.6792 0.9700 0.9849
0.4162 14.6316 556 0.9221 0.6573 0.9221 0.9603
0.4162 14.6842 558 0.8751 0.6423 0.8751 0.9355
0.4162 14.7368 560 0.8957 0.6522 0.8957 0.9464

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
2
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_FineTuningAraBERT_run2_AugV5_k8_task1_organization

Finetuned
(4023)
this model