ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k8_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7439
  • Qwk: 0.4785
  • Mse: 0.7439
  • Rmse: 0.8625

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0385 2 4.0176 -0.0169 4.0176 2.0044
No log 0.0769 4 1.9728 0.0727 1.9728 1.4046
No log 0.1154 6 1.2187 0.0696 1.2187 1.1039
No log 0.1538 8 1.6201 0.0487 1.6201 1.2728
No log 0.1923 10 1.1827 0.0089 1.1827 1.0875
No log 0.2308 12 0.8342 -0.0205 0.8342 0.9133
No log 0.2692 14 0.7428 0.0417 0.7428 0.8618
No log 0.3077 16 0.7751 0.0143 0.7751 0.8804
No log 0.3462 18 1.0092 0.0254 1.0092 1.0046
No log 0.3846 20 1.2481 -0.0067 1.2481 1.1172
No log 0.4231 22 1.3872 -0.0010 1.3872 1.1778
No log 0.4615 24 1.1300 0.1101 1.1300 1.0630
No log 0.5 26 0.7467 0.2443 0.7467 0.8641
No log 0.5385 28 0.7163 0.2445 0.7163 0.8463
No log 0.5769 30 0.8286 0.2646 0.8286 0.9103
No log 0.6154 32 0.9760 0.1702 0.9760 0.9879
No log 0.6538 34 1.2220 0.1131 1.2220 1.1055
No log 0.6923 36 1.2550 0.1620 1.2550 1.1203
No log 0.7308 38 0.9984 0.2055 0.9984 0.9992
No log 0.7692 40 0.9815 0.1635 0.9815 0.9907
No log 0.8077 42 1.1830 0.2235 1.1830 1.0876
No log 0.8462 44 1.3151 0.2398 1.3151 1.1468
No log 0.8846 46 1.6001 0.1647 1.6001 1.2650
No log 0.9231 48 1.6718 0.1573 1.6718 1.2930
No log 0.9615 50 1.5582 0.1438 1.5582 1.2483
No log 1.0 52 1.2423 0.1125 1.2423 1.1146
No log 1.0385 54 0.8639 0.2560 0.8639 0.9295
No log 1.0769 56 0.7861 0.3285 0.7861 0.8866
No log 1.1154 58 0.8470 0.2165 0.8470 0.9203
No log 1.1538 60 0.9363 0.2129 0.9363 0.9676
No log 1.1923 62 1.0343 0.1823 1.0343 1.0170
No log 1.2308 64 1.2409 0.1350 1.2409 1.1140
No log 1.2692 66 1.2843 0.1386 1.2843 1.1333
No log 1.3077 68 1.4141 0.1471 1.4141 1.1892
No log 1.3462 70 1.1420 0.1580 1.1420 1.0687
No log 1.3846 72 0.9561 0.2676 0.9561 0.9778
No log 1.4231 74 0.8571 0.2837 0.8571 0.9258
No log 1.4615 76 0.7493 0.3769 0.7493 0.8656
No log 1.5 78 0.6639 0.3997 0.6639 0.8148
No log 1.5385 80 0.7175 0.3575 0.7175 0.8470
No log 1.5769 82 0.7290 0.2933 0.7290 0.8538
No log 1.6154 84 0.8278 0.2063 0.8278 0.9098
No log 1.6538 86 0.9172 0.2295 0.9172 0.9577
No log 1.6923 88 1.1240 0.1949 1.1240 1.0602
No log 1.7308 90 1.1519 0.1708 1.1519 1.0733
No log 1.7692 92 1.0165 0.2200 1.0165 1.0082
No log 1.8077 94 0.8209 0.2874 0.8209 0.9060
No log 1.8462 96 0.6899 0.4069 0.6899 0.8306
No log 1.8846 98 0.6670 0.3998 0.6670 0.8167
No log 1.9231 100 0.6467 0.4041 0.6467 0.8041
No log 1.9615 102 0.6439 0.3751 0.6439 0.8024
No log 2.0 104 0.5712 0.4318 0.5712 0.7558
No log 2.0385 106 0.5325 0.4751 0.5325 0.7298
No log 2.0769 108 0.5112 0.5081 0.5112 0.7150
No log 2.1154 110 0.5132 0.4824 0.5132 0.7164
No log 2.1538 112 0.5479 0.4490 0.5479 0.7402
No log 2.1923 114 0.7271 0.3670 0.7271 0.8527
No log 2.2308 116 1.0129 0.2808 1.0129 1.0064
No log 2.2692 118 1.1586 0.2320 1.1586 1.0764
No log 2.3077 120 0.9531 0.2858 0.9531 0.9763
No log 2.3462 122 0.8558 0.3530 0.8558 0.9251
No log 2.3846 124 0.8570 0.3772 0.8570 0.9257
No log 2.4231 126 0.8178 0.3832 0.8178 0.9043
No log 2.4615 128 0.7951 0.4108 0.7951 0.8917
No log 2.5 130 0.8736 0.4021 0.8736 0.9347
No log 2.5385 132 1.0213 0.3656 1.0213 1.0106
No log 2.5769 134 1.0397 0.3364 1.0397 1.0197
No log 2.6154 136 0.8710 0.3904 0.8710 0.9333
No log 2.6538 138 0.7027 0.4262 0.7027 0.8383
No log 2.6923 140 0.7143 0.3591 0.7143 0.8451
No log 2.7308 142 0.8089 0.3131 0.8089 0.8994
No log 2.7692 144 0.8101 0.3314 0.8101 0.9001
No log 2.8077 146 0.7410 0.3593 0.7410 0.8608
No log 2.8462 148 0.6896 0.4509 0.6896 0.8304
No log 2.8846 150 0.7155 0.4769 0.7155 0.8459
No log 2.9231 152 0.7069 0.4439 0.7069 0.8408
No log 2.9615 154 0.6939 0.4257 0.6939 0.8330
No log 3.0 156 0.7350 0.3975 0.7350 0.8573
No log 3.0385 158 0.7294 0.4186 0.7294 0.8541
No log 3.0769 160 0.6908 0.4195 0.6908 0.8311
No log 3.1154 162 0.7351 0.4475 0.7351 0.8574
No log 3.1538 164 0.8311 0.4371 0.8311 0.9116
No log 3.1923 166 0.8099 0.4514 0.8099 0.8999
No log 3.2308 168 0.7295 0.4689 0.7295 0.8541
No log 3.2692 170 0.7208 0.4667 0.7208 0.8490
No log 3.3077 172 0.7615 0.4427 0.7615 0.8726
No log 3.3462 174 0.7832 0.4540 0.7832 0.8850
No log 3.3846 176 0.7466 0.4432 0.7466 0.8641
No log 3.4231 178 0.7160 0.4913 0.7160 0.8462
No log 3.4615 180 0.6971 0.4505 0.6971 0.8349
No log 3.5 182 0.7018 0.4285 0.7018 0.8377
No log 3.5385 184 0.7259 0.4330 0.7259 0.8520
No log 3.5769 186 0.7757 0.4457 0.7757 0.8808
No log 3.6154 188 0.8193 0.4268 0.8193 0.9052
No log 3.6538 190 0.7891 0.4049 0.7891 0.8883
No log 3.6923 192 0.7706 0.4576 0.7706 0.8778
No log 3.7308 194 0.7707 0.4464 0.7707 0.8779
No log 3.7692 196 0.7898 0.4719 0.7898 0.8887
No log 3.8077 198 0.7750 0.4678 0.7750 0.8803
No log 3.8462 200 0.7016 0.4534 0.7016 0.8376
No log 3.8846 202 0.6623 0.4617 0.6623 0.8138
No log 3.9231 204 0.6606 0.4493 0.6606 0.8128
No log 3.9615 206 0.6897 0.4470 0.6897 0.8305
No log 4.0 208 0.7023 0.4493 0.7023 0.8381
No log 4.0385 210 0.6917 0.4242 0.6917 0.8317
No log 4.0769 212 0.6878 0.4087 0.6878 0.8293
No log 4.1154 214 0.7004 0.4242 0.7004 0.8369
No log 4.1538 216 0.7454 0.4613 0.7454 0.8634
No log 4.1923 218 0.7740 0.4629 0.7740 0.8798
No log 4.2308 220 0.7934 0.4835 0.7934 0.8907
No log 4.2692 222 0.7662 0.4389 0.7662 0.8753
No log 4.3077 224 0.7296 0.4514 0.7296 0.8542
No log 4.3462 226 0.7234 0.5154 0.7234 0.8505
No log 4.3846 228 0.7381 0.4643 0.7381 0.8591
No log 4.4231 230 0.7337 0.4863 0.7337 0.8565
No log 4.4615 232 0.6908 0.5173 0.6908 0.8312
No log 4.5 234 0.6984 0.4606 0.6984 0.8357
No log 4.5385 236 0.6989 0.4670 0.6989 0.8360
No log 4.5769 238 0.7047 0.4976 0.7047 0.8395
No log 4.6154 240 0.7809 0.5195 0.7809 0.8837
No log 4.6538 242 0.8093 0.4927 0.8093 0.8996
No log 4.6923 244 0.7579 0.4684 0.7579 0.8706
No log 4.7308 246 0.7337 0.4458 0.7337 0.8566
No log 4.7692 248 0.7430 0.4495 0.7430 0.8620
No log 4.8077 250 0.7680 0.5062 0.7680 0.8764
No log 4.8462 252 0.7852 0.5127 0.7852 0.8861
No log 4.8846 254 0.7702 0.4864 0.7702 0.8776
No log 4.9231 256 0.7029 0.4909 0.7029 0.8384
No log 4.9615 258 0.6673 0.4605 0.6673 0.8169
No log 5.0 260 0.6693 0.4966 0.6693 0.8181
No log 5.0385 262 0.6825 0.5069 0.6825 0.8262
No log 5.0769 264 0.6825 0.5148 0.6825 0.8262
No log 5.1154 266 0.6698 0.4776 0.6698 0.8184
No log 5.1538 268 0.6764 0.4741 0.6764 0.8225
No log 5.1923 270 0.6943 0.5208 0.6943 0.8332
No log 5.2308 272 0.7070 0.4955 0.7070 0.8408
No log 5.2692 274 0.7043 0.5214 0.7043 0.8392
No log 5.3077 276 0.7052 0.5048 0.7052 0.8397
No log 5.3462 278 0.7116 0.5174 0.7116 0.8436
No log 5.3846 280 0.7236 0.5098 0.7236 0.8507
No log 5.4231 282 0.7258 0.5038 0.7258 0.8520
No log 5.4615 284 0.7260 0.4584 0.7260 0.8520
No log 5.5 286 0.7373 0.4524 0.7373 0.8586
No log 5.5385 288 0.7146 0.4700 0.7146 0.8454
No log 5.5769 290 0.6982 0.4537 0.6982 0.8356
No log 5.6154 292 0.7156 0.4692 0.7156 0.8459
No log 5.6538 294 0.7412 0.5169 0.7412 0.8609
No log 5.6923 296 0.7318 0.4957 0.7318 0.8554
No log 5.7308 298 0.6996 0.4595 0.6996 0.8364
No log 5.7692 300 0.6899 0.4776 0.6899 0.8306
No log 5.8077 302 0.7378 0.4213 0.7378 0.8589
No log 5.8462 304 0.7738 0.4153 0.7738 0.8797
No log 5.8846 306 0.7493 0.4135 0.7493 0.8656
No log 5.9231 308 0.7072 0.4554 0.7072 0.8409
No log 5.9615 310 0.6803 0.4984 0.6803 0.8248
No log 6.0 312 0.6915 0.4731 0.6915 0.8316
No log 6.0385 314 0.6922 0.4792 0.6922 0.8320
No log 6.0769 316 0.6880 0.5012 0.6880 0.8294
No log 6.1154 318 0.6949 0.5003 0.6949 0.8336
No log 6.1538 320 0.7054 0.4716 0.7054 0.8399
No log 6.1923 322 0.7143 0.4670 0.7143 0.8452
No log 6.2308 324 0.7159 0.4417 0.7159 0.8461
No log 6.2692 326 0.7248 0.4659 0.7248 0.8513
No log 6.3077 328 0.7214 0.4687 0.7214 0.8494
No log 6.3462 330 0.7056 0.4659 0.7056 0.8400
No log 6.3846 332 0.6894 0.4747 0.6894 0.8303
No log 6.4231 334 0.6844 0.4747 0.6844 0.8273
No log 6.4615 336 0.6898 0.4507 0.6898 0.8305
No log 6.5 338 0.6903 0.4847 0.6903 0.8309
No log 6.5385 340 0.6905 0.4596 0.6905 0.8310
No log 6.5769 342 0.7047 0.4497 0.7047 0.8395
No log 6.6154 344 0.7119 0.4497 0.7119 0.8438
No log 6.6538 346 0.7168 0.4497 0.7168 0.8466
No log 6.6923 348 0.7184 0.4632 0.7184 0.8476
No log 6.7308 350 0.7114 0.4553 0.7114 0.8434
No log 6.7692 352 0.6991 0.4641 0.6991 0.8361
No log 6.8077 354 0.6883 0.4814 0.6883 0.8296
No log 6.8462 356 0.6845 0.5012 0.6845 0.8274
No log 6.8846 358 0.6808 0.5108 0.6808 0.8251
No log 6.9231 360 0.6840 0.5108 0.6840 0.8271
No log 6.9615 362 0.6967 0.4822 0.6967 0.8347
No log 7.0 364 0.7030 0.4745 0.7030 0.8385
No log 7.0385 366 0.7058 0.4745 0.7058 0.8401
No log 7.0769 368 0.6967 0.4782 0.6967 0.8347
No log 7.1154 370 0.6976 0.4877 0.6976 0.8352
No log 7.1538 372 0.7001 0.4705 0.7001 0.8367
No log 7.1923 374 0.7031 0.4763 0.7031 0.8385
No log 7.2308 376 0.7057 0.4852 0.7057 0.8400
No log 7.2692 378 0.7145 0.4895 0.7145 0.8453
No log 7.3077 380 0.7302 0.5058 0.7302 0.8545
No log 7.3462 382 0.7354 0.4785 0.7354 0.8576
No log 7.3846 384 0.7286 0.4870 0.7286 0.8536
No log 7.4231 386 0.7307 0.4785 0.7307 0.8548
No log 7.4615 388 0.7198 0.4739 0.7198 0.8484
No log 7.5 390 0.7153 0.4662 0.7153 0.8458
No log 7.5385 392 0.7110 0.4662 0.7110 0.8432
No log 7.5769 394 0.7035 0.4739 0.7035 0.8387
No log 7.6154 396 0.6923 0.4646 0.6923 0.8321
No log 7.6538 398 0.6883 0.4825 0.6883 0.8296
No log 7.6923 400 0.6853 0.4970 0.6853 0.8278
No log 7.7308 402 0.6845 0.5275 0.6845 0.8273
No log 7.7692 404 0.6788 0.5118 0.6788 0.8239
No log 7.8077 406 0.6768 0.5064 0.6768 0.8227
No log 7.8462 408 0.6796 0.4985 0.6796 0.8244
No log 7.8846 410 0.6958 0.4925 0.6958 0.8342
No log 7.9231 412 0.7170 0.5023 0.7170 0.8467
No log 7.9615 414 0.7185 0.5023 0.7185 0.8476
No log 8.0 416 0.7063 0.4785 0.7063 0.8404
No log 8.0385 418 0.6909 0.5064 0.6909 0.8312
No log 8.0769 420 0.6885 0.5021 0.6885 0.8298
No log 8.1154 422 0.6885 0.5021 0.6885 0.8298
No log 8.1538 424 0.6906 0.5021 0.6906 0.8310
No log 8.1923 426 0.6975 0.5147 0.6975 0.8352
No log 8.2308 428 0.7051 0.4845 0.7051 0.8397
No log 8.2692 430 0.7120 0.4803 0.7120 0.8438
No log 8.3077 432 0.7149 0.4791 0.7149 0.8455
No log 8.3462 434 0.7115 0.4707 0.7115 0.8435
No log 8.3846 436 0.7102 0.4707 0.7102 0.8427
No log 8.4231 438 0.7091 0.4782 0.7091 0.8421
No log 8.4615 440 0.7093 0.4725 0.7093 0.8422
No log 8.5 442 0.7092 0.5088 0.7092 0.8421
No log 8.5385 444 0.7093 0.5088 0.7093 0.8422
No log 8.5769 446 0.7114 0.5070 0.7114 0.8434
No log 8.6154 448 0.7164 0.4752 0.7164 0.8464
No log 8.6538 450 0.7224 0.4585 0.7224 0.8499
No log 8.6923 452 0.7257 0.4710 0.7257 0.8519
No log 8.7308 454 0.7276 0.4710 0.7276 0.8530
No log 8.7692 456 0.7289 0.4710 0.7289 0.8538
No log 8.8077 458 0.7239 0.4581 0.7239 0.8508
No log 8.8462 460 0.7259 0.4581 0.7259 0.8520
No log 8.8846 462 0.7268 0.4710 0.7268 0.8525
No log 8.9231 464 0.7216 0.4581 0.7216 0.8495
No log 8.9615 466 0.7225 0.4562 0.7225 0.8500
No log 9.0 468 0.7275 0.4562 0.7275 0.8529
No log 9.0385 470 0.7385 0.4686 0.7385 0.8593
No log 9.0769 472 0.7487 0.4593 0.7487 0.8653
No log 9.1154 474 0.7523 0.4650 0.7523 0.8674
No log 9.1538 476 0.7505 0.4593 0.7505 0.8663
No log 9.1923 478 0.7505 0.4593 0.7505 0.8663
No log 9.2308 480 0.7467 0.4874 0.7467 0.8641
No log 9.2692 482 0.7482 0.4686 0.7482 0.8650
No log 9.3077 484 0.7481 0.4686 0.7481 0.8649
No log 9.3462 486 0.7445 0.4691 0.7445 0.8628
No log 9.3846 488 0.7441 0.4767 0.7441 0.8626
No log 9.4231 490 0.7473 0.4767 0.7473 0.8644
No log 9.4615 492 0.7529 0.4767 0.7529 0.8677
No log 9.5 494 0.7607 0.4552 0.7607 0.8722
No log 9.5385 496 0.7655 0.4552 0.7655 0.8749
No log 9.5769 498 0.7679 0.4552 0.7679 0.8763
0.4351 9.6154 500 0.7669 0.4552 0.7669 0.8757
0.4351 9.6538 502 0.7636 0.4552 0.7636 0.8738
0.4351 9.6923 504 0.7597 0.4630 0.7597 0.8716
0.4351 9.7308 506 0.7562 0.4630 0.7562 0.8696
0.4351 9.7692 508 0.7526 0.4710 0.7526 0.8675
0.4351 9.8077 510 0.7494 0.4785 0.7494 0.8657
0.4351 9.8462 512 0.7469 0.4785 0.7469 0.8642
0.4351 9.8846 514 0.7456 0.4785 0.7456 0.8635
0.4351 9.9231 516 0.7446 0.4785 0.7446 0.8629
0.4351 9.9615 518 0.7441 0.4785 0.7441 0.8626
0.4351 10.0 520 0.7439 0.4785 0.7439 0.8625

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run2_AugV5_k8_task2_organization

Finetuned
(4023)
this model