ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k16_task2_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.0191
  • Qwk: 0.3404
  • Mse: 1.0191
  • Rmse: 1.0095

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0253 2 4.3669 0.0010 4.3669 2.0897
No log 0.0506 4 2.4413 -0.0118 2.4413 1.5625
No log 0.0759 6 1.7479 0.0372 1.7479 1.3221
No log 0.1013 8 1.3135 -0.0009 1.3135 1.1461
No log 0.1266 10 1.1580 0.1944 1.1580 1.0761
No log 0.1519 12 1.1516 0.2537 1.1516 1.0731
No log 0.1772 14 1.5196 0.1222 1.5196 1.2327
No log 0.2025 16 1.4430 0.1169 1.4430 1.2012
No log 0.2278 18 1.1163 0.2351 1.1163 1.0565
No log 0.2532 20 1.1578 0.1974 1.1578 1.0760
No log 0.2785 22 1.3916 0.0994 1.3916 1.1797
No log 0.3038 24 1.3427 0.0936 1.3427 1.1588
No log 0.3291 26 1.4319 0.1418 1.4319 1.1966
No log 0.3544 28 1.2325 0.2304 1.2325 1.1102
No log 0.3797 30 1.2132 0.1974 1.2132 1.1015
No log 0.4051 32 1.0485 0.2395 1.0485 1.0240
No log 0.4304 34 1.0374 0.2738 1.0374 1.0185
No log 0.4557 36 1.1514 0.2203 1.1514 1.0730
No log 0.4810 38 1.5541 0.1805 1.5541 1.2466
No log 0.5063 40 1.5524 0.1805 1.5524 1.2460
No log 0.5316 42 1.2510 0.1984 1.2510 1.1185
No log 0.5570 44 1.1954 0.2763 1.1954 1.0933
No log 0.5823 46 1.2232 0.2983 1.2232 1.1060
No log 0.6076 48 1.1003 0.3323 1.1003 1.0490
No log 0.6329 50 1.1521 0.3092 1.1521 1.0734
No log 0.6582 52 1.1530 0.3323 1.1530 1.0738
No log 0.6835 54 1.2909 0.2578 1.2909 1.1362
No log 0.7089 56 1.2810 0.2072 1.2810 1.1318
No log 0.7342 58 1.1193 0.3307 1.1193 1.0580
No log 0.7595 60 1.1166 0.3584 1.1166 1.0567
No log 0.7848 62 0.9656 0.4101 0.9656 0.9826
No log 0.8101 64 0.7787 0.5987 0.7787 0.8824
No log 0.8354 66 0.7884 0.5607 0.7884 0.8879
No log 0.8608 68 1.0994 0.3949 1.0994 1.0485
No log 0.8861 70 1.7673 0.3116 1.7673 1.3294
No log 0.9114 72 1.9306 0.2122 1.9306 1.3895
No log 0.9367 74 1.8235 0.2590 1.8235 1.3504
No log 0.9620 76 1.4695 0.3701 1.4695 1.2122
No log 0.9873 78 1.1431 0.3363 1.1431 1.0691
No log 1.0127 80 1.0906 0.3339 1.0906 1.0443
No log 1.0380 82 1.2369 0.3275 1.2369 1.1122
No log 1.0633 84 1.5105 0.3233 1.5105 1.2290
No log 1.0886 86 1.8259 0.3026 1.8259 1.3513
No log 1.1139 88 1.9500 0.2757 1.9500 1.3964
No log 1.1392 90 1.6097 0.3368 1.6097 1.2687
No log 1.1646 92 1.2707 0.4346 1.2707 1.1272
No log 1.1899 94 1.4170 0.3934 1.4170 1.1904
No log 1.2152 96 1.3310 0.3972 1.3310 1.1537
No log 1.2405 98 1.3166 0.3972 1.3166 1.1474
No log 1.2658 100 1.0806 0.4721 1.0806 1.0395
No log 1.2911 102 0.9406 0.5420 0.9406 0.9698
No log 1.3165 104 0.9649 0.5307 0.9649 0.9823
No log 1.3418 106 0.9901 0.5039 0.9901 0.9951
No log 1.3671 108 0.9574 0.4685 0.9574 0.9785
No log 1.3924 110 0.9559 0.5606 0.9559 0.9777
No log 1.4177 112 1.0282 0.3980 1.0282 1.0140
No log 1.4430 114 1.0374 0.4280 1.0374 1.0185
No log 1.4684 116 1.0052 0.4640 1.0052 1.0026
No log 1.4937 118 0.9794 0.4731 0.9794 0.9896
No log 1.5190 120 0.9666 0.4812 0.9666 0.9831
No log 1.5443 122 1.1198 0.3922 1.1198 1.0582
No log 1.5696 124 1.2081 0.3719 1.2081 1.0991
No log 1.5949 126 1.0588 0.4431 1.0588 1.0290
No log 1.6203 128 1.0061 0.4549 1.0061 1.0030
No log 1.6456 130 0.9702 0.5255 0.9702 0.9850
No log 1.6709 132 0.9244 0.4385 0.9244 0.9615
No log 1.6962 134 0.9537 0.4615 0.9537 0.9766
No log 1.7215 136 0.9968 0.3887 0.9968 0.9984
No log 1.7468 138 0.9069 0.4715 0.9069 0.9523
No log 1.7722 140 0.8879 0.4832 0.8879 0.9423
No log 1.7975 142 0.8781 0.4930 0.8781 0.9371
No log 1.8228 144 0.9366 0.4364 0.9366 0.9678
No log 1.8481 146 0.9761 0.3625 0.9761 0.9880
No log 1.8734 148 0.9183 0.4316 0.9183 0.9583
No log 1.8987 150 0.9390 0.4963 0.9390 0.9690
No log 1.9241 152 0.9731 0.4449 0.9731 0.9865
No log 1.9494 154 1.2983 0.3517 1.2983 1.1394
No log 1.9747 156 1.3019 0.3630 1.3019 1.1410
No log 2.0 158 1.1090 0.4308 1.1090 1.0531
No log 2.0253 160 1.1082 0.4387 1.1082 1.0527
No log 2.0506 162 0.9893 0.4428 0.9893 0.9946
No log 2.0759 164 1.0230 0.4970 1.0230 1.0114
No log 2.1013 166 1.1349 0.4465 1.1349 1.0653
No log 2.1266 168 1.0351 0.4031 1.0351 1.0174
No log 2.1519 170 0.9626 0.4728 0.9626 0.9811
No log 2.1772 172 1.1185 0.4318 1.1185 1.0576
No log 2.2025 174 1.1458 0.3711 1.1458 1.0704
No log 2.2278 176 0.9763 0.4140 0.9763 0.9881
No log 2.2532 178 0.9433 0.4595 0.9433 0.9712
No log 2.2785 180 1.0988 0.4167 1.0988 1.0482
No log 2.3038 182 1.2148 0.2857 1.2148 1.1022
No log 2.3291 184 1.1298 0.3830 1.1298 1.0629
No log 2.3544 186 1.0024 0.4540 1.0024 1.0012
No log 2.3797 188 1.1049 0.4157 1.1049 1.0512
No log 2.4051 190 1.4510 0.2867 1.4510 1.2046
No log 2.4304 192 1.4803 0.2612 1.4803 1.2167
No log 2.4557 194 1.1735 0.4458 1.1735 1.0833
No log 2.4810 196 1.0698 0.4624 1.0698 1.0343
No log 2.5063 198 1.0895 0.4481 1.0895 1.0438
No log 2.5316 200 1.2080 0.4489 1.2080 1.0991
No log 2.5570 202 1.2683 0.3550 1.2683 1.1262
No log 2.5823 204 1.1578 0.4289 1.1578 1.0760
No log 2.6076 206 1.0137 0.4508 1.0137 1.0068
No log 2.6329 208 0.9986 0.4601 0.9986 0.9993
No log 2.6582 210 1.0297 0.4066 1.0297 1.0147
No log 2.6835 212 1.0296 0.4420 1.0296 1.0147
No log 2.7089 214 1.0464 0.4103 1.0464 1.0230
No log 2.7342 216 1.0230 0.4728 1.0230 1.0114
No log 2.7595 218 1.0262 0.4055 1.0262 1.0130
No log 2.7848 220 1.0255 0.4055 1.0255 1.0127
No log 2.8101 222 1.0712 0.4210 1.0712 1.0350
No log 2.8354 224 1.1191 0.4282 1.1191 1.0579
No log 2.8608 226 1.1408 0.4014 1.1408 1.0681
No log 2.8861 228 1.0537 0.4335 1.0537 1.0265
No log 2.9114 230 0.9898 0.4145 0.9898 0.9949
No log 2.9367 232 1.0349 0.3990 1.0349 1.0173
No log 2.9620 234 1.0425 0.4200 1.0425 1.0210
No log 2.9873 236 0.9644 0.5233 0.9644 0.9821
No log 3.0127 238 1.0606 0.3584 1.0606 1.0299
No log 3.0380 240 1.3119 0.3061 1.3119 1.1454
No log 3.0633 242 1.3148 0.2705 1.3148 1.1466
No log 3.0886 244 1.0911 0.3226 1.0911 1.0446
No log 3.1139 246 0.9252 0.4257 0.9252 0.9619
No log 3.1392 248 0.9318 0.4832 0.9318 0.9653
No log 3.1646 250 0.9159 0.4219 0.9159 0.9570
No log 3.1899 252 0.9497 0.3104 0.9497 0.9745
No log 3.2152 254 0.9682 0.3231 0.9682 0.9840
No log 3.2405 256 1.0428 0.3584 1.0428 1.0212
No log 3.2658 258 1.0212 0.3330 1.0212 1.0105
No log 3.2911 260 0.9187 0.5085 0.9187 0.9585
No log 3.3165 262 0.9098 0.5232 0.9098 0.9538
No log 3.3418 264 0.9284 0.5054 0.9284 0.9635
No log 3.3671 266 1.0353 0.4107 1.0353 1.0175
No log 3.3924 268 1.0129 0.4848 1.0129 1.0065
No log 3.4177 270 0.9757 0.4964 0.9757 0.9878
No log 3.4430 272 0.9627 0.5213 0.9627 0.9812
No log 3.4684 274 0.9659 0.4373 0.9659 0.9828
No log 3.4937 276 0.9726 0.3596 0.9726 0.9862
No log 3.5190 278 1.0547 0.3149 1.0547 1.0270
No log 3.5443 280 1.1104 0.2870 1.1104 1.0538
No log 3.5696 282 1.0478 0.2844 1.0478 1.0236
No log 3.5949 284 0.9790 0.3695 0.9790 0.9894
No log 3.6203 286 0.9853 0.4048 0.9853 0.9926
No log 3.6456 288 0.9900 0.3445 0.9900 0.9950
No log 3.6709 290 1.0495 0.4962 1.0495 1.0245
No log 3.6962 292 1.0079 0.4211 1.0079 1.0039
No log 3.7215 294 0.9429 0.3943 0.9429 0.9710
No log 3.7468 296 0.9409 0.3596 0.9409 0.9700
No log 3.7722 298 0.9787 0.4045 0.9787 0.9893
No log 3.7975 300 1.0272 0.4101 1.0272 1.0135
No log 3.8228 302 1.0869 0.4172 1.0869 1.0426
No log 3.8481 304 1.0590 0.4088 1.0590 1.0291
No log 3.8734 306 0.9691 0.4243 0.9691 0.9844
No log 3.8987 308 0.9301 0.4232 0.9301 0.9644
No log 3.9241 310 0.9040 0.4081 0.9040 0.9508
No log 3.9494 312 0.9212 0.4476 0.9212 0.9598
No log 3.9747 314 0.9213 0.4242 0.9213 0.9598
No log 4.0 316 0.9374 0.3159 0.9374 0.9682
No log 4.0253 318 1.1558 0.3523 1.1558 1.0751
No log 4.0506 320 1.3876 0.3171 1.3876 1.1779
No log 4.0759 322 1.2920 0.3434 1.2920 1.1367
No log 4.1013 324 1.1087 0.2896 1.1087 1.0529
No log 4.1266 326 1.0098 0.4470 1.0098 1.0049
No log 4.1519 328 1.1035 0.4 1.1035 1.0505
No log 4.1772 330 1.1090 0.4200 1.1090 1.0531
No log 4.2025 332 1.0483 0.4337 1.0483 1.0239
No log 4.2278 334 1.0555 0.2103 1.0555 1.0274
No log 4.2532 336 1.1487 0.1990 1.1487 1.0718
No log 4.2785 338 1.2372 0.2890 1.2372 1.1123
No log 4.3038 340 1.3088 0.2554 1.3088 1.1440
No log 4.3291 342 1.2237 0.2890 1.2237 1.1062
No log 4.3544 344 1.0861 0.2075 1.0861 1.0422
No log 4.3797 346 1.0582 0.2075 1.0582 1.0287
No log 4.4051 348 1.0571 0.2311 1.0571 1.0282
No log 4.4304 350 0.9867 0.3015 0.9867 0.9933
No log 4.4557 352 0.9667 0.3164 0.9667 0.9832
No log 4.4810 354 0.9715 0.3861 0.9715 0.9857
No log 4.5063 356 0.9689 0.3164 0.9689 0.9843
No log 4.5316 358 0.9656 0.3113 0.9656 0.9826
No log 4.5570 360 0.9571 0.3090 0.9571 0.9783
No log 4.5823 362 0.9695 0.3472 0.9695 0.9846
No log 4.6076 364 1.0027 0.3287 1.0027 1.0013
No log 4.6329 366 1.0359 0.3968 1.0359 1.0178
No log 4.6582 368 1.0369 0.3972 1.0369 1.0183
No log 4.6835 370 0.9971 0.3126 0.9971 0.9985
No log 4.7089 372 0.9678 0.3891 0.9678 0.9837
No log 4.7342 374 0.9810 0.3584 0.9810 0.9904
No log 4.7595 376 1.0533 0.3298 1.0533 1.0263
No log 4.7848 378 1.0011 0.2939 1.0011 1.0006
No log 4.8101 380 0.9682 0.3336 0.9682 0.9840
No log 4.8354 382 0.9873 0.3286 0.9873 0.9936
No log 4.8608 384 1.0524 0.2942 1.0524 1.0259
No log 4.8861 386 1.0558 0.3305 1.0558 1.0275
No log 4.9114 388 1.0940 0.3389 1.0940 1.0459
No log 4.9367 390 1.0306 0.2942 1.0306 1.0152
No log 4.9620 392 1.0171 0.3321 1.0171 1.0085
No log 4.9873 394 0.9650 0.2897 0.9650 0.9823
No log 5.0127 396 0.9506 0.4045 0.9506 0.9750
No log 5.0380 398 0.9527 0.4181 0.9527 0.9760
No log 5.0633 400 1.0236 0.3705 1.0236 1.0117
No log 5.0886 402 1.1199 0.3430 1.1199 1.0582
No log 5.1139 404 1.0821 0.3521 1.0821 1.0403
No log 5.1392 406 0.9740 0.375 0.9740 0.9869
No log 5.1646 408 0.9592 0.3908 0.9592 0.9794
No log 5.1899 410 0.9601 0.4104 0.9601 0.9798
No log 5.2152 412 0.9545 0.3164 0.9545 0.9770
No log 5.2405 414 0.9680 0.2991 0.9680 0.9839
No log 5.2658 416 0.9995 0.3657 0.9995 0.9997
No log 5.2911 418 0.9837 0.4212 0.9837 0.9918
No log 5.3165 420 0.9683 0.4119 0.9683 0.9840
No log 5.3418 422 0.9823 0.4156 0.9823 0.9911
No log 5.3671 424 0.9516 0.4783 0.9516 0.9755
No log 5.3924 426 0.9139 0.4739 0.9139 0.9560
No log 5.4177 428 0.9111 0.4304 0.9111 0.9545
No log 5.4430 430 0.9054 0.5026 0.9054 0.9515
No log 5.4684 432 0.9014 0.3991 0.9014 0.9494
No log 5.4937 434 0.9085 0.3399 0.9085 0.9531
No log 5.5190 436 0.9235 0.3542 0.9235 0.9610
No log 5.5443 438 0.9255 0.3542 0.9255 0.9620
No log 5.5696 440 0.9074 0.3541 0.9074 0.9526
No log 5.5949 442 0.8920 0.3541 0.8920 0.9445
No log 5.6203 444 0.8801 0.3695 0.8801 0.9381
No log 5.6456 446 0.8724 0.4548 0.8724 0.9340
No log 5.6709 448 0.8657 0.5011 0.8657 0.9304
No log 5.6962 450 0.8613 0.5431 0.8613 0.9280
No log 5.7215 452 0.8660 0.4853 0.8660 0.9306
No log 5.7468 454 0.8694 0.4643 0.8694 0.9324
No log 5.7722 456 0.8897 0.4045 0.8897 0.9433
No log 5.7975 458 0.9081 0.3989 0.9081 0.9530
No log 5.8228 460 0.8865 0.4292 0.8865 0.9415
No log 5.8481 462 0.8864 0.4122 0.8864 0.9415
No log 5.8734 464 0.8964 0.4122 0.8964 0.9468
No log 5.8987 466 0.9237 0.3908 0.9237 0.9611
No log 5.9241 468 0.9809 0.3663 0.9809 0.9904
No log 5.9494 470 0.9980 0.3935 0.9980 0.9990
No log 5.9747 472 0.9665 0.3059 0.9665 0.9831
No log 6.0 474 0.9516 0.3608 0.9516 0.9755
No log 6.0253 476 0.9515 0.3327 0.9515 0.9754
No log 6.0506 478 0.9567 0.2919 0.9567 0.9781
No log 6.0759 480 0.9896 0.3126 0.9896 0.9948
No log 6.1013 482 0.9938 0.3614 0.9938 0.9969
No log 6.1266 484 0.9851 0.3243 0.9851 0.9925
No log 6.1519 486 0.9761 0.3243 0.9761 0.9880
No log 6.1772 488 0.9598 0.3059 0.9598 0.9797
No log 6.2025 490 0.9346 0.3363 0.9346 0.9667
No log 6.2278 492 0.9256 0.3363 0.9256 0.9621
No log 6.2532 494 0.9098 0.3363 0.9098 0.9538
No log 6.2785 496 0.9013 0.3692 0.9013 0.9494
No log 6.3038 498 0.8969 0.3788 0.8969 0.9471
0.3208 6.3291 500 0.9047 0.3555 0.9047 0.9511
0.3208 6.3544 502 0.9068 0.3555 0.9068 0.9523
0.3208 6.3797 504 0.9078 0.3555 0.9078 0.9528
0.3208 6.4051 506 0.9382 0.4045 0.9382 0.9686
0.3208 6.4304 508 0.9507 0.3625 0.9507 0.9751
0.3208 6.4557 510 0.9280 0.3346 0.9280 0.9633
0.3208 6.4810 512 0.9243 0.3458 0.9243 0.9614
0.3208 6.5063 514 0.9321 0.3497 0.9321 0.9654
0.3208 6.5316 516 0.9598 0.3294 0.9598 0.9797
0.3208 6.5570 518 1.0191 0.3404 1.0191 1.0095

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingALLEssays_FineTuningAraBERT_run3_AugV5_k16_task2_organization

Finetuned
(4019)
this model