ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k13_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8840
  • Qwk: 0.4139
  • Mse: 0.8840
  • Rmse: 0.9402

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0426 2 4.6435 0.0010 4.6435 2.1549
No log 0.0851 4 2.6726 -0.0233 2.6726 1.6348
No log 0.1277 6 2.0031 0.0770 2.0031 1.4153
No log 0.1702 8 1.6472 0.0062 1.6472 1.2834
No log 0.2128 10 1.6149 0.0585 1.6149 1.2708
No log 0.2553 12 1.9987 0.1174 1.9987 1.4138
No log 0.2979 14 2.0027 0.1054 2.0027 1.4152
No log 0.3404 16 1.5901 0.0749 1.5901 1.2610
No log 0.3830 18 1.2793 0.1257 1.2793 1.1311
No log 0.4255 20 1.1880 0.2386 1.1880 1.0900
No log 0.4681 22 1.2379 0.1865 1.2379 1.1126
No log 0.5106 24 1.1940 0.1999 1.1940 1.0927
No log 0.5532 26 1.2968 0.1882 1.2968 1.1388
No log 0.5957 28 1.5793 0.0254 1.5793 1.2567
No log 0.6383 30 1.6659 0.0772 1.6659 1.2907
No log 0.6809 32 1.3236 0.1495 1.3236 1.1505
No log 0.7234 34 1.0196 0.3243 1.0196 1.0098
No log 0.7660 36 0.9862 0.3189 0.9862 0.9931
No log 0.8085 38 1.2444 0.0898 1.2444 1.1155
No log 0.8511 40 1.9671 0.2092 1.9671 1.4025
No log 0.8936 42 2.0903 0.2390 2.0903 1.4458
No log 0.9362 44 1.8408 0.1898 1.8408 1.3568
No log 0.9787 46 1.4361 0.1522 1.4361 1.1984
No log 1.0213 48 1.2171 0.1570 1.2171 1.1032
No log 1.0638 50 1.1151 0.2543 1.1151 1.0560
No log 1.1064 52 1.0568 0.3106 1.0568 1.0280
No log 1.1489 54 1.0117 0.2961 1.0117 1.0058
No log 1.1915 56 1.1132 0.2191 1.1132 1.0551
No log 1.2340 58 1.2798 0.2815 1.2798 1.1313
No log 1.2766 60 1.6785 0.2505 1.6785 1.2956
No log 1.3191 62 1.7247 0.2619 1.7247 1.3133
No log 1.3617 64 1.5733 0.2630 1.5733 1.2543
No log 1.4043 66 1.4176 0.2210 1.4176 1.1906
No log 1.4468 68 1.0280 0.3113 1.0280 1.0139
No log 1.4894 70 0.9250 0.4115 0.9250 0.9618
No log 1.5319 72 0.9095 0.3699 0.9095 0.9537
No log 1.5745 74 1.1611 0.3629 1.1611 1.0776
No log 1.6170 76 1.4779 0.2524 1.4779 1.2157
No log 1.6596 78 1.5966 0.2877 1.5966 1.2636
No log 1.7021 80 1.2012 0.3978 1.2012 1.0960
No log 1.7447 82 1.0684 0.2532 1.0684 1.0336
No log 1.7872 84 1.1843 0.2723 1.1843 1.0883
No log 1.8298 86 1.2739 0.1761 1.2739 1.1287
No log 1.8723 88 1.0508 0.3052 1.0508 1.0251
No log 1.9149 90 0.9354 0.3633 0.9354 0.9672
No log 1.9574 92 0.9861 0.3959 0.9861 0.9930
No log 2.0 94 0.8698 0.3886 0.8698 0.9326
No log 2.0426 96 0.8068 0.4273 0.8068 0.8982
No log 2.0851 98 0.6905 0.5962 0.6905 0.8310
No log 2.1277 100 0.7670 0.6011 0.7670 0.8758
No log 2.1702 102 0.6960 0.6035 0.6960 0.8343
No log 2.2128 104 0.6627 0.5342 0.6627 0.8140
No log 2.2553 106 0.6745 0.4912 0.6745 0.8213
No log 2.2979 108 0.6732 0.5120 0.6732 0.8205
No log 2.3404 110 0.8315 0.5111 0.8315 0.9119
No log 2.3830 112 1.2516 0.2411 1.2516 1.1187
No log 2.4255 114 1.1776 0.2259 1.1776 1.0852
No log 2.4681 116 0.7840 0.5362 0.7840 0.8854
No log 2.5106 118 1.0290 0.5165 1.0290 1.0144
No log 2.5532 120 1.3717 0.3053 1.3717 1.1712
No log 2.5957 122 1.1981 0.3966 1.1981 1.0946
No log 2.6383 124 0.8608 0.3956 0.8608 0.9278
No log 2.6809 126 0.8157 0.4488 0.8157 0.9032
No log 2.7234 128 0.9592 0.5780 0.9592 0.9794
No log 2.7660 130 0.9121 0.5799 0.9121 0.9550
No log 2.8085 132 0.7893 0.4343 0.7893 0.8884
No log 2.8511 134 0.8128 0.4737 0.8128 0.9015
No log 2.8936 136 0.9604 0.3798 0.9604 0.9800
No log 2.9362 138 0.9147 0.4371 0.9147 0.9564
No log 2.9787 140 0.8024 0.4591 0.8024 0.8958
No log 3.0213 142 0.7925 0.4799 0.7925 0.8902
No log 3.0638 144 0.7917 0.4998 0.7917 0.8898
No log 3.1064 146 0.7838 0.5061 0.7838 0.8853
No log 3.1489 148 1.0855 0.4138 1.0855 1.0418
No log 3.1915 150 1.5935 0.3942 1.5935 1.2624
No log 3.2340 152 1.6151 0.3851 1.6151 1.2709
No log 3.2766 154 1.3776 0.4152 1.3776 1.1737
No log 3.3191 156 0.9042 0.5055 0.9042 0.9509
No log 3.3617 158 0.7453 0.5339 0.7453 0.8633
No log 3.4043 160 0.8812 0.5124 0.8812 0.9387
No log 3.4468 162 0.8223 0.5266 0.8223 0.9068
No log 3.4894 164 0.7382 0.4859 0.7382 0.8592
No log 3.5319 166 0.7710 0.5125 0.7710 0.8780
No log 3.5745 168 0.7970 0.5125 0.7970 0.8928
No log 3.6170 170 0.7847 0.5012 0.7847 0.8858
No log 3.6596 172 0.8684 0.5179 0.8684 0.9319
No log 3.7021 174 0.9806 0.4613 0.9806 0.9903
No log 3.7447 176 0.9266 0.4650 0.9266 0.9626
No log 3.7872 178 0.8097 0.4138 0.8097 0.8999
No log 3.8298 180 0.8175 0.3660 0.8175 0.9041
No log 3.8723 182 0.8465 0.4400 0.8465 0.9201
No log 3.9149 184 0.8391 0.4885 0.8391 0.9160
No log 3.9574 186 0.8414 0.5477 0.8414 0.9173
No log 4.0 188 0.7995 0.5816 0.7995 0.8941
No log 4.0426 190 0.7908 0.6176 0.7908 0.8893
No log 4.0851 192 0.7893 0.5622 0.7893 0.8884
No log 4.1277 194 0.8379 0.4952 0.8379 0.9153
No log 4.1702 196 0.9282 0.4681 0.9282 0.9634
No log 4.2128 198 0.8703 0.4966 0.8703 0.9329
No log 4.2553 200 0.7429 0.4979 0.7429 0.8619
No log 4.2979 202 0.7797 0.4570 0.7797 0.8830
No log 4.3404 204 0.8299 0.4350 0.8299 0.9110
No log 4.3830 206 0.7787 0.4175 0.7787 0.8824
No log 4.4255 208 0.7571 0.4571 0.7571 0.8701
No log 4.4681 210 0.8332 0.4845 0.8332 0.9128
No log 4.5106 212 0.8356 0.5261 0.8356 0.9141
No log 4.5532 214 0.7708 0.5345 0.7708 0.8779
No log 4.5957 216 0.7413 0.5708 0.7413 0.8610
No log 4.6383 218 0.7397 0.5550 0.7397 0.8601
No log 4.6809 220 0.7326 0.6324 0.7326 0.8559
No log 4.7234 222 0.7356 0.6118 0.7356 0.8577
No log 4.7660 224 0.7498 0.5555 0.7498 0.8659
No log 4.8085 226 0.7606 0.5555 0.7606 0.8721
No log 4.8511 228 0.7859 0.5726 0.7859 0.8865
No log 4.8936 230 0.8257 0.5447 0.8257 0.9087
No log 4.9362 232 0.8671 0.4023 0.8671 0.9312
No log 4.9787 234 0.8332 0.5216 0.8332 0.9128
No log 5.0213 236 0.8343 0.4278 0.8343 0.9134
No log 5.0638 238 0.8569 0.4340 0.8569 0.9257
No log 5.1064 240 0.8922 0.3914 0.8922 0.9446
No log 5.1489 242 0.9076 0.3814 0.9076 0.9527
No log 5.1915 244 0.9254 0.3759 0.9254 0.9620
No log 5.2340 246 1.0408 0.4410 1.0408 1.0202
No log 5.2766 248 1.0429 0.4383 1.0429 1.0212
No log 5.3191 250 0.8878 0.4344 0.8878 0.9422
No log 5.3617 252 0.8365 0.4982 0.8365 0.9146
No log 5.4043 254 0.9921 0.4558 0.9921 0.9960
No log 5.4468 256 1.0740 0.4625 1.0740 1.0363
No log 5.4894 258 0.9014 0.4949 0.9014 0.9494
No log 5.5319 260 0.7813 0.5186 0.7813 0.8839
No log 5.5745 262 0.9196 0.4494 0.9196 0.9590
No log 5.6170 264 0.9657 0.5057 0.9657 0.9827
No log 5.6596 266 0.8154 0.5210 0.8154 0.9030
No log 5.7021 268 0.7503 0.5708 0.7503 0.8662
No log 5.7447 270 0.8191 0.5514 0.8191 0.9050
No log 5.7872 272 0.8513 0.5391 0.8513 0.9227
No log 5.8298 274 0.7978 0.5130 0.7978 0.8932
No log 5.8723 276 0.7534 0.4908 0.7534 0.8680
No log 5.9149 278 0.7723 0.4958 0.7723 0.8788
No log 5.9574 280 0.7468 0.4676 0.7468 0.8642
No log 6.0 282 0.7766 0.3812 0.7766 0.8812
No log 6.0426 284 0.8661 0.4640 0.8661 0.9307
No log 6.0851 286 0.8808 0.5080 0.8808 0.9385
No log 6.1277 288 0.8103 0.4912 0.8103 0.9002
No log 6.1702 290 0.7678 0.4817 0.7678 0.8762
No log 6.2128 292 0.7582 0.5046 0.7582 0.8707
No log 6.2553 294 0.7685 0.5163 0.7685 0.8766
No log 6.2979 296 0.7668 0.5606 0.7668 0.8757
No log 6.3404 298 0.7525 0.6133 0.7525 0.8675
No log 6.3830 300 0.7464 0.5879 0.7464 0.8639
No log 6.4255 302 0.7357 0.6043 0.7357 0.8577
No log 6.4681 304 0.7447 0.4118 0.7447 0.8630
No log 6.5106 306 0.7317 0.4016 0.7317 0.8554
No log 6.5532 308 0.7065 0.5150 0.7065 0.8406
No log 6.5957 310 0.7239 0.6235 0.7239 0.8508
No log 6.6383 312 0.7811 0.5047 0.7811 0.8838
No log 6.6809 314 0.8341 0.4138 0.8341 0.9133
No log 6.7234 316 0.9108 0.3648 0.9108 0.9544
No log 6.7660 318 0.8839 0.3648 0.8839 0.9402
No log 6.8085 320 0.8738 0.3859 0.8738 0.9348
No log 6.8511 322 0.8597 0.4139 0.8597 0.9272
No log 6.8936 324 0.8389 0.4389 0.8389 0.9159
No log 6.9362 326 0.8074 0.5416 0.8074 0.8986
No log 6.9787 328 0.7725 0.5416 0.7725 0.8789
No log 7.0213 330 0.7923 0.5339 0.7923 0.8901
No log 7.0638 332 0.7647 0.5820 0.7647 0.8745
No log 7.1064 334 0.7466 0.5770 0.7466 0.8641
No log 7.1489 336 0.7648 0.5312 0.7648 0.8745
No log 7.1915 338 0.7861 0.5076 0.7861 0.8866
No log 7.2340 340 0.8139 0.4995 0.8139 0.9022
No log 7.2766 342 0.8915 0.5309 0.8915 0.9442
No log 7.3191 344 0.8933 0.5309 0.8933 0.9452
No log 7.3617 346 0.8488 0.5047 0.8488 0.9213
No log 7.4043 348 0.8271 0.3762 0.8271 0.9095
No log 7.4468 350 0.8429 0.3855 0.8429 0.9181
No log 7.4894 352 0.8067 0.3914 0.8067 0.8982
No log 7.5319 354 0.8057 0.5047 0.8057 0.8976
No log 7.5745 356 0.9529 0.4596 0.9529 0.9762
No log 7.6170 358 1.0201 0.4284 1.0201 1.0100
No log 7.6596 360 0.9619 0.4596 0.9619 0.9808
No log 7.7021 362 0.8686 0.4203 0.8686 0.9320
No log 7.7447 364 0.8484 0.4016 0.8484 0.9211
No log 7.7872 366 0.8625 0.3486 0.8625 0.9287
No log 7.8298 368 0.8497 0.3821 0.8497 0.9218
No log 7.8723 370 0.8453 0.5202 0.8453 0.9194
No log 7.9149 372 0.8993 0.4365 0.8993 0.9483
No log 7.9574 374 0.9099 0.4225 0.9099 0.9539
No log 8.0 376 0.9065 0.3879 0.9065 0.9521
No log 8.0426 378 0.8996 0.3812 0.8996 0.9485
No log 8.0851 380 0.8738 0.3847 0.8738 0.9348
No log 8.1277 382 0.8551 0.4571 0.8551 0.9247
No log 8.1702 384 0.8540 0.4036 0.8540 0.9241
No log 8.2128 386 0.8511 0.4282 0.8511 0.9225
No log 8.2553 388 0.8583 0.4324 0.8583 0.9264
No log 8.2979 390 0.8817 0.3747 0.8817 0.9390
No log 8.3404 392 0.9120 0.3646 0.9120 0.9550
No log 8.3830 394 0.9115 0.3590 0.9115 0.9547
No log 8.4255 396 0.9063 0.4180 0.9063 0.9520
No log 8.4681 398 0.9037 0.3779 0.9037 0.9506
No log 8.5106 400 0.8961 0.4428 0.8961 0.9466
No log 8.5532 402 0.8849 0.3866 0.8849 0.9407
No log 8.5957 404 0.8714 0.4016 0.8714 0.9335
No log 8.6383 406 0.8575 0.3862 0.8575 0.9260
No log 8.6809 408 0.8589 0.3572 0.8589 0.9268
No log 8.7234 410 0.8319 0.4271 0.8319 0.9121
No log 8.7660 412 0.7958 0.4860 0.7958 0.8921
No log 8.8085 414 0.8075 0.4488 0.8075 0.8986
No log 8.8511 416 0.8461 0.4402 0.8461 0.9199
No log 8.8936 418 0.8277 0.4488 0.8277 0.9098
No log 8.9362 420 0.8112 0.3974 0.8112 0.9007
No log 8.9787 422 0.8530 0.4036 0.8530 0.9236
No log 9.0213 424 0.8988 0.3602 0.8988 0.9480
No log 9.0638 426 0.8685 0.3699 0.8685 0.9319
No log 9.1064 428 0.8185 0.4401 0.8185 0.9047
No log 9.1489 430 0.8170 0.4180 0.8170 0.9039
No log 9.1915 432 0.8738 0.4265 0.8738 0.9348
No log 9.2340 434 0.8892 0.4965 0.8892 0.9430
No log 9.2766 436 0.8271 0.5161 0.8271 0.9095
No log 9.3191 438 0.7755 0.5606 0.7755 0.8806
No log 9.3617 440 0.7834 0.4749 0.7834 0.8851
No log 9.4043 442 0.8070 0.4741 0.8070 0.8983
No log 9.4468 444 0.8013 0.4494 0.8013 0.8951
No log 9.4894 446 0.8105 0.4494 0.8105 0.9003
No log 9.5319 448 0.8214 0.3756 0.8214 0.9063
No log 9.5745 450 0.8242 0.4299 0.8242 0.9079
No log 9.6170 452 0.8412 0.3855 0.8412 0.9172
No log 9.6596 454 0.8646 0.4136 0.8646 0.9298
No log 9.7021 456 0.8842 0.3709 0.8842 0.9403
No log 9.7447 458 0.8716 0.3678 0.8716 0.9336
No log 9.7872 460 0.8384 0.4715 0.8384 0.9156
No log 9.8298 462 0.8427 0.5421 0.8427 0.9180
No log 9.8723 464 0.8684 0.5270 0.8684 0.9319
No log 9.9149 466 0.8443 0.5421 0.8443 0.9189
No log 9.9574 468 0.8443 0.4548 0.8443 0.9188
No log 10.0 470 0.9643 0.3444 0.9643 0.9820
No log 10.0426 472 1.0702 0.3301 1.0702 1.0345
No log 10.0851 474 1.0366 0.3160 1.0366 1.0181
No log 10.1277 476 0.9168 0.3699 0.9168 0.9575
No log 10.1702 478 0.8475 0.3830 0.8475 0.9206
No log 10.2128 480 0.8682 0.5014 0.8682 0.9317
No log 10.2553 482 0.8723 0.4742 0.8723 0.9340
No log 10.2979 484 0.8724 0.4640 0.8724 0.9340
No log 10.3404 486 0.8418 0.4163 0.8418 0.9175
No log 10.3830 488 0.8098 0.5112 0.8098 0.8999
No log 10.4255 490 0.8027 0.5175 0.8027 0.8960
No log 10.4681 492 0.8060 0.5076 0.8060 0.8977
No log 10.5106 494 0.8119 0.4722 0.8119 0.9011
No log 10.5532 496 0.8673 0.4640 0.8673 0.9313
No log 10.5957 498 0.8649 0.4640 0.8649 0.9300
0.3429 10.6383 500 0.8340 0.4243 0.8340 0.9132
0.3429 10.6809 502 0.8426 0.4261 0.8426 0.9179
0.3429 10.7234 504 0.8610 0.3602 0.8610 0.9279
0.3429 10.7660 506 0.8391 0.3806 0.8391 0.9160
0.3429 10.8085 508 0.8175 0.4701 0.8175 0.9042
0.3429 10.8511 510 0.8184 0.4571 0.8184 0.9046
0.3429 10.8936 512 0.8195 0.3467 0.8195 0.9053
0.3429 10.9362 514 0.8257 0.3671 0.8257 0.9087
0.3429 10.9787 516 0.8460 0.4236 0.8460 0.9198
0.3429 11.0213 518 0.8493 0.4408 0.8493 0.9215
0.3429 11.0638 520 0.8581 0.4408 0.8581 0.9264
0.3429 11.1064 522 0.8631 0.4136 0.8631 0.9290
0.3429 11.1489 524 0.8536 0.3365 0.8536 0.9239
0.3429 11.1915 526 0.8621 0.3557 0.8621 0.9285
0.3429 11.2340 528 0.8728 0.3927 0.8728 0.9342
0.3429 11.2766 530 0.8840 0.4139 0.8840 0.9402

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k13_task2_organization

Finetuned
(4019)
this model