ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0293
  • Qwk: 0.3090
  • Mse: 1.0293
  • Rmse: 1.0145

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0690 2 4.6106 0.0010 4.6106 2.1472
No log 0.1379 4 2.7680 0.0104 2.7680 1.6637
No log 0.2069 6 2.2589 -0.0846 2.2589 1.5030
No log 0.2759 8 1.6807 0.0116 1.6807 1.2964
No log 0.3448 10 1.2881 0.1043 1.2881 1.1350
No log 0.4138 12 1.1804 0.1379 1.1804 1.0865
No log 0.4828 14 1.2802 0.1753 1.2802 1.1315
No log 0.5517 16 1.2380 0.1320 1.2380 1.1127
No log 0.6207 18 1.2008 0.2137 1.2008 1.0958
No log 0.6897 20 1.1867 0.2095 1.1867 1.0893
No log 0.7586 22 1.1533 0.2584 1.1533 1.0739
No log 0.8276 24 1.1112 0.2936 1.1112 1.0541
No log 0.8966 26 1.1641 0.2395 1.1641 1.0790
No log 0.9655 28 1.3359 0.1076 1.3359 1.1558
No log 1.0345 30 1.3844 0.0426 1.3844 1.1766
No log 1.1034 32 1.4400 0.0426 1.4400 1.2000
No log 1.1724 34 1.4359 0.0562 1.4359 1.1983
No log 1.2414 36 1.3973 0.0426 1.3973 1.1821
No log 1.3103 38 1.2834 0.0838 1.2834 1.1329
No log 1.3793 40 1.1968 0.0955 1.1968 1.0940
No log 1.4483 42 1.1755 0.0955 1.1755 1.0842
No log 1.5172 44 1.2946 0.1226 1.2946 1.1378
No log 1.5862 46 1.3311 0.1051 1.3311 1.1538
No log 1.6552 48 1.1679 0.1636 1.1679 1.0807
No log 1.7241 50 1.0226 0.4748 1.0226 1.0113
No log 1.7931 52 1.1199 0.1722 1.1199 1.0582
No log 1.8621 54 1.3274 0.0361 1.3274 1.1521
No log 1.9310 56 1.1270 0.1732 1.1270 1.0616
No log 2.0 58 1.0434 0.3693 1.0434 1.0215
No log 2.0690 60 1.2653 0.1584 1.2653 1.1249
No log 2.1379 62 1.2947 0.1257 1.2947 1.1378
No log 2.2069 64 1.1047 0.2609 1.1047 1.0510
No log 2.2759 66 0.9714 0.4958 0.9714 0.9856
No log 2.3448 68 0.9590 0.4180 0.9590 0.9793
No log 2.4138 70 0.9431 0.5485 0.9431 0.9711
No log 2.4828 72 1.0831 0.2870 1.0831 1.0407
No log 2.5517 74 1.3125 0.2399 1.3125 1.1457
No log 2.6207 76 1.2478 0.2815 1.2478 1.1171
No log 2.6897 78 0.9811 0.3624 0.9811 0.9905
No log 2.7586 80 0.9820 0.3976 0.9820 0.9909
No log 2.8276 82 0.9505 0.4032 0.9505 0.9749
No log 2.8966 84 1.0700 0.3395 1.0700 1.0344
No log 2.9655 86 1.0199 0.3602 1.0199 1.0099
No log 3.0345 88 0.8742 0.4294 0.8742 0.9350
No log 3.1034 90 0.8407 0.4889 0.8407 0.9169
No log 3.1724 92 0.8303 0.5029 0.8303 0.9112
No log 3.2414 94 0.8457 0.5059 0.8457 0.9196
No log 3.3103 96 0.9453 0.4741 0.9453 0.9723
No log 3.3793 98 1.3273 0.3405 1.3273 1.1521
No log 3.4483 100 1.3719 0.3158 1.3719 1.1713
No log 3.5172 102 1.0238 0.3892 1.0238 1.0118
No log 3.5862 104 0.8352 0.4603 0.8352 0.9139
No log 3.6552 106 0.9983 0.3705 0.9983 0.9992
No log 3.7241 108 1.0975 0.3838 1.0975 1.0476
No log 3.7931 110 0.9368 0.4938 0.9368 0.9679
No log 3.8621 112 0.9107 0.4491 0.9107 0.9543
No log 3.9310 114 1.0006 0.3179 1.0006 1.0003
No log 4.0 116 1.0004 0.3276 1.0004 1.0002
No log 4.0690 118 0.8783 0.4555 0.8783 0.9372
No log 4.1379 120 0.8866 0.4775 0.8866 0.9416
No log 4.2069 122 0.8721 0.4832 0.8721 0.9338
No log 4.2759 124 0.8657 0.4637 0.8657 0.9304
No log 4.3448 126 0.9728 0.4522 0.9728 0.9863
No log 4.4138 128 1.0878 0.4318 1.0878 1.0430
No log 4.4828 130 0.9733 0.4752 0.9733 0.9866
No log 4.5517 132 0.8812 0.5443 0.8812 0.9387
No log 4.6207 134 0.8908 0.4503 0.8908 0.9438
No log 4.6897 136 0.8850 0.5426 0.8850 0.9408
No log 4.7586 138 0.9452 0.5593 0.9452 0.9722
No log 4.8276 140 0.9483 0.5423 0.9483 0.9738
No log 4.8966 142 0.8507 0.4724 0.8507 0.9223
No log 4.9655 144 0.9503 0.4167 0.9503 0.9748
No log 5.0345 146 1.0441 0.3800 1.0441 1.0218
No log 5.1034 148 0.9545 0.4037 0.9545 0.9770
No log 5.1724 150 0.8875 0.4595 0.8875 0.9421
No log 5.2414 152 0.8496 0.5223 0.8496 0.9217
No log 5.3103 154 0.8426 0.4996 0.8426 0.9179
No log 5.3793 156 0.8727 0.5057 0.8727 0.9342
No log 5.4483 158 0.9394 0.4200 0.9394 0.9692
No log 5.5172 160 0.9095 0.4270 0.9095 0.9537
No log 5.5862 162 0.8414 0.5137 0.8414 0.9173
No log 5.6552 164 0.8613 0.4479 0.8613 0.9281
No log 5.7241 166 0.9012 0.4540 0.9012 0.9493
No log 5.7931 168 0.9025 0.3845 0.9025 0.9500
No log 5.8621 170 0.9168 0.3845 0.9168 0.9575
No log 5.9310 172 0.9653 0.4529 0.9653 0.9825
No log 6.0 174 0.9876 0.4529 0.9876 0.9938
No log 6.0690 176 0.9654 0.3960 0.9654 0.9825
No log 6.1379 178 0.9313 0.4609 0.9313 0.9650
No log 6.2069 180 0.9116 0.4593 0.9116 0.9548
No log 6.2759 182 0.8970 0.4122 0.8970 0.9471
No log 6.3448 184 0.9315 0.4690 0.9315 0.9651
No log 6.4138 186 0.9660 0.4876 0.9660 0.9828
No log 6.4828 188 0.8802 0.4143 0.8802 0.9382
No log 6.5517 190 0.8664 0.4313 0.8664 0.9308
No log 6.6207 192 0.8915 0.4796 0.8915 0.9442
No log 6.6897 194 0.8776 0.4724 0.8776 0.9368
No log 6.7586 196 0.8770 0.4671 0.8770 0.9365
No log 6.8276 198 0.9010 0.4393 0.9010 0.9492
No log 6.8966 200 0.8978 0.4280 0.8978 0.9475
No log 6.9655 202 0.9333 0.3720 0.9333 0.9661
No log 7.0345 204 0.9367 0.3957 0.9367 0.9678
No log 7.1034 206 0.9236 0.3392 0.9236 0.9610
No log 7.1724 208 0.9219 0.3695 0.9219 0.9602
No log 7.2414 210 0.9304 0.4002 0.9304 0.9646
No log 7.3103 212 1.0084 0.4838 1.0084 1.0042
No log 7.3793 214 1.0096 0.4935 1.0096 1.0048
No log 7.4483 216 0.9538 0.4334 0.9538 0.9766
No log 7.5172 218 0.9390 0.3943 0.9390 0.9690
No log 7.5862 220 0.9433 0.3554 0.9433 0.9712
No log 7.6552 222 0.9266 0.4045 0.9266 0.9626
No log 7.7241 224 0.9322 0.4603 0.9322 0.9655
No log 7.7931 226 0.9573 0.4202 0.9573 0.9784
No log 7.8621 228 0.9470 0.3961 0.9470 0.9731
No log 7.9310 230 0.9533 0.4221 0.9533 0.9764
No log 8.0 232 1.0698 0.2702 1.0698 1.0343
No log 8.0690 234 1.1201 0.2844 1.1201 1.0584
No log 8.1379 236 1.0268 0.2702 1.0268 1.0133
No log 8.2069 238 0.9400 0.4681 0.9400 0.9695
No log 8.2759 240 0.9348 0.4275 0.9348 0.9668
No log 8.3448 242 0.9344 0.4930 0.9344 0.9666
No log 8.4138 244 0.9924 0.3859 0.9924 0.9962
No log 8.4828 246 0.9872 0.3767 0.9872 0.9936
No log 8.5517 248 0.9268 0.4521 0.9268 0.9627
No log 8.6207 250 0.9407 0.4459 0.9407 0.9699
No log 8.6897 252 0.9746 0.4420 0.9746 0.9872
No log 8.7586 254 0.9499 0.4666 0.9499 0.9746
No log 8.8276 256 0.9141 0.4203 0.9141 0.9561
No log 8.8966 258 0.8907 0.4444 0.8907 0.9438
No log 8.9655 260 0.8861 0.4203 0.8861 0.9413
No log 9.0345 262 0.8735 0.4767 0.8735 0.9346
No log 9.1034 264 0.8869 0.4916 0.8869 0.9418
No log 9.1724 266 0.8700 0.5165 0.8700 0.9327
No log 9.2414 268 0.8546 0.4541 0.8546 0.9245
No log 9.3103 270 0.8705 0.5244 0.8705 0.9330
No log 9.3793 272 0.8397 0.4541 0.8397 0.9164
No log 9.4483 274 0.8423 0.5169 0.8423 0.9178
No log 9.5172 276 0.8718 0.5098 0.8718 0.9337
No log 9.5862 278 0.8485 0.4681 0.8485 0.9211
No log 9.6552 280 0.8446 0.4978 0.8446 0.9190
No log 9.7241 282 0.8727 0.4916 0.8727 0.9342
No log 9.7931 284 0.8641 0.4792 0.8641 0.9295
No log 9.8621 286 0.8483 0.4285 0.8483 0.9210
No log 9.9310 288 0.8659 0.4757 0.8659 0.9306
No log 10.0 290 0.8679 0.4699 0.8679 0.9316
No log 10.0690 292 0.8725 0.4285 0.8725 0.9341
No log 10.1379 294 0.8841 0.4343 0.8841 0.9403
No log 10.2069 296 0.8977 0.4275 0.8977 0.9475
No log 10.2759 298 0.9030 0.4714 0.9030 0.9503
No log 10.3448 300 0.8936 0.4411 0.8936 0.9453
No log 10.4138 302 0.9274 0.4050 0.9274 0.9630
No log 10.4828 304 0.9694 0.4655 0.9694 0.9846
No log 10.5517 306 0.9355 0.4299 0.9355 0.9672
No log 10.6207 308 0.8997 0.4106 0.8997 0.9485
No log 10.6897 310 0.9022 0.4035 0.9022 0.9498
No log 10.7586 312 0.9057 0.4377 0.9057 0.9517
No log 10.8276 314 0.9111 0.4318 0.9111 0.9545
No log 10.8966 316 0.9168 0.3983 0.9168 0.9575
No log 10.9655 318 0.9064 0.4318 0.9064 0.9521
No log 11.0345 320 0.8980 0.4634 0.8980 0.9476
No log 11.1034 322 0.8876 0.4161 0.8876 0.9421
No log 11.1724 324 0.8973 0.4435 0.8973 0.9473
No log 11.2414 326 0.8971 0.4124 0.8971 0.9472
No log 11.3103 328 0.9061 0.4411 0.9061 0.9519
No log 11.3793 330 0.9262 0.4667 0.9262 0.9624
No log 11.4483 332 0.9145 0.3766 0.9145 0.9563
No log 11.5172 334 0.9268 0.4690 0.9268 0.9627
No log 11.5862 336 0.9721 0.4966 0.9721 0.9859
No log 11.6552 338 0.9616 0.4532 0.9616 0.9806
No log 11.7241 340 0.9430 0.3621 0.9430 0.9711
No log 11.7931 342 0.9393 0.4042 0.9393 0.9692
No log 11.8621 344 0.9436 0.3661 0.9436 0.9714
No log 11.9310 346 0.9603 0.3688 0.9603 0.9799
No log 12.0 348 0.9325 0.4290 0.9325 0.9657
No log 12.0690 350 0.9116 0.4609 0.9116 0.9548
No log 12.1379 352 0.8948 0.4254 0.8948 0.9459
No log 12.2069 354 0.8900 0.4254 0.8900 0.9434
No log 12.2759 356 0.9005 0.4667 0.9005 0.9490
No log 12.3448 358 0.8904 0.3977 0.8904 0.9436
No log 12.4138 360 0.8927 0.4087 0.8927 0.9448
No log 12.4828 362 0.9158 0.3577 0.9158 0.9570
No log 12.5517 364 0.9290 0.3577 0.9290 0.9639
No log 12.6207 366 0.9326 0.3948 0.9326 0.9657
No log 12.6897 368 0.9436 0.4349 0.9436 0.9714
No log 12.7586 370 0.9701 0.3819 0.9701 0.9849
No log 12.8276 372 0.9989 0.3677 0.9989 0.9995
No log 12.8966 374 0.9892 0.3478 0.9892 0.9946
No log 12.9655 376 0.9807 0.3607 0.9807 0.9903
No log 13.0345 378 1.0103 0.4120 1.0103 1.0051
No log 13.1034 380 1.0262 0.4104 1.0262 1.0130
No log 13.1724 382 0.9721 0.4090 0.9721 0.9859
No log 13.2414 384 0.9392 0.4554 0.9392 0.9691
No log 13.3103 386 0.9451 0.4767 0.9451 0.9721
No log 13.3793 388 0.9407 0.4032 0.9407 0.9699
No log 13.4483 390 0.9426 0.3539 0.9426 0.9709
No log 13.5172 392 0.9413 0.3886 0.9413 0.9702
No log 13.5862 394 0.9262 0.3486 0.9262 0.9624
No log 13.6552 396 0.9185 0.3764 0.9185 0.9584
No log 13.7241 398 0.9165 0.3652 0.9165 0.9574
No log 13.7931 400 0.9111 0.3652 0.9111 0.9545
No log 13.8621 402 0.9053 0.3424 0.9053 0.9515
No log 13.9310 404 0.9133 0.4420 0.9133 0.9557
No log 14.0 406 0.9479 0.3956 0.9479 0.9736
No log 14.0690 408 0.9543 0.3728 0.9543 0.9769
No log 14.1379 410 0.9132 0.4197 0.9132 0.9556
No log 14.2069 412 0.8976 0.3868 0.8976 0.9474
No log 14.2759 414 0.9146 0.4851 0.9146 0.9563
No log 14.3448 416 0.9076 0.4434 0.9076 0.9527
No log 14.4138 418 0.8944 0.4002 0.8944 0.9458
No log 14.4828 420 0.9094 0.4019 0.9094 0.9536
No log 14.5517 422 0.9249 0.3681 0.9249 0.9617
No log 14.6207 424 0.9314 0.3066 0.9314 0.9651
No log 14.6897 426 0.9256 0.3392 0.9256 0.9621
No log 14.7586 428 0.9235 0.3493 0.9235 0.9610
No log 14.8276 430 0.9141 0.3493 0.9141 0.9561
No log 14.8966 432 0.9031 0.3796 0.9031 0.9503
No log 14.9655 434 0.8929 0.4026 0.8929 0.9449
No log 15.0345 436 0.8831 0.4420 0.8831 0.9397
No log 15.1034 438 0.8756 0.4420 0.8756 0.9357
No log 15.1724 440 0.8723 0.4124 0.8723 0.9340
No log 15.2414 442 0.8780 0.4181 0.8780 0.9370
No log 15.3103 444 0.8906 0.4470 0.8906 0.9437
No log 15.3793 446 0.8935 0.4006 0.8935 0.9453
No log 15.4483 448 0.8898 0.3796 0.8898 0.9433
No log 15.5172 450 0.9224 0.4268 0.9224 0.9604
No log 15.5862 452 0.9509 0.3544 0.9509 0.9751
No log 15.6552 454 0.9258 0.3775 0.9258 0.9622
No log 15.7241 456 0.8831 0.4220 0.8831 0.9397
No log 15.7931 458 0.8738 0.4521 0.8738 0.9348
No log 15.8621 460 0.8748 0.4898 0.8748 0.9353
No log 15.9310 462 0.8858 0.4898 0.8858 0.9412
No log 16.0 464 0.8741 0.4617 0.8741 0.9349
No log 16.0690 466 0.8776 0.3478 0.8776 0.9368
No log 16.1379 468 0.9014 0.4398 0.9014 0.9494
No log 16.2069 470 0.9009 0.4241 0.9009 0.9492
No log 16.2759 472 0.9061 0.4079 0.9061 0.9519
No log 16.3448 474 0.9252 0.4061 0.9252 0.9619
No log 16.4138 476 0.9226 0.4159 0.9226 0.9605
No log 16.4828 478 0.9037 0.4180 0.9037 0.9507
No log 16.5517 480 0.8888 0.4377 0.8888 0.9428
No log 16.6207 482 0.8859 0.4241 0.8859 0.9412
No log 16.6897 484 0.8902 0.4318 0.8902 0.9435
No log 16.7586 486 0.8977 0.4292 0.8977 0.9475
No log 16.8276 488 0.8892 0.4450 0.8892 0.9430
No log 16.8966 490 0.8825 0.4313 0.8825 0.9394
No log 16.9655 492 0.8795 0.3914 0.8795 0.9378
No log 17.0345 494 0.8859 0.3966 0.8859 0.9412
No log 17.1034 496 0.8827 0.3217 0.8827 0.9395
No log 17.1724 498 0.8856 0.4257 0.8856 0.9411
0.3038 17.2414 500 0.9260 0.4556 0.9260 0.9623
0.3038 17.3103 502 0.9294 0.4549 0.9294 0.9640
0.3038 17.3793 504 0.8974 0.4617 0.8974 0.9473
0.3038 17.4483 506 0.8977 0.3621 0.8977 0.9475
0.3038 17.5172 508 0.9108 0.4202 0.9108 0.9543
0.3038 17.5862 510 0.9094 0.4002 0.9094 0.9536
0.3038 17.6552 512 0.9251 0.4356 0.9251 0.9618
0.3038 17.7241 514 1.0293 0.3090 1.0293 1.0145

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k6_task2_organization

Finetuned
(4019)
this model