ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k14_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.6808
  • Qwk: 0.4622
  • Mse: 0.6808
  • Rmse: 0.8251

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0282 2 4.0522 0.0086 4.0522 2.0130
No log 0.0563 4 2.2135 -0.0443 2.2135 1.4878
No log 0.0845 6 1.3723 0.0232 1.3723 1.1715
No log 0.1127 8 1.1944 0.2293 1.1944 1.0929
No log 0.1408 10 1.1628 0.1891 1.1628 1.0783
No log 0.1690 12 1.1746 0.1740 1.1746 1.0838
No log 0.1972 14 1.1098 0.2114 1.1098 1.0535
No log 0.2254 16 1.0546 0.1657 1.0546 1.0269
No log 0.2535 18 1.0984 0.2208 1.0984 1.0480
No log 0.2817 20 1.8090 0.1437 1.8090 1.3450
No log 0.3099 22 2.0178 0.1401 2.0178 1.4205
No log 0.3380 24 1.5771 0.1706 1.5771 1.2558
No log 0.3662 26 1.1730 0.1352 1.1730 1.0830
No log 0.3944 28 1.0649 0.2125 1.0649 1.0319
No log 0.4225 30 1.0268 0.2246 1.0268 1.0133
No log 0.4507 32 1.0516 0.2125 1.0516 1.0255
No log 0.4789 34 1.0120 0.1832 1.0120 1.0060
No log 0.5070 36 0.8906 0.4148 0.8906 0.9437
No log 0.5352 38 0.8799 0.4692 0.8799 0.9380
No log 0.5634 40 0.8988 0.4312 0.8988 0.9481
No log 0.5915 42 0.8985 0.5274 0.8985 0.9479
No log 0.6197 44 1.0046 0.3317 1.0046 1.0023
No log 0.6479 46 1.0633 0.3569 1.0633 1.0312
No log 0.6761 48 0.9800 0.4620 0.9800 0.9899
No log 0.7042 50 1.0837 0.4177 1.0837 1.0410
No log 0.7324 52 1.0447 0.4694 1.0447 1.0221
No log 0.7606 54 0.8984 0.3643 0.8984 0.9479
No log 0.7887 56 1.0607 0.2752 1.0607 1.0299
No log 0.8169 58 0.9495 0.3908 0.9495 0.9744
No log 0.8451 60 0.8402 0.4663 0.8402 0.9166
No log 0.8732 62 0.9063 0.5316 0.9063 0.9520
No log 0.9014 64 0.9686 0.5362 0.9686 0.9842
No log 0.9296 66 0.8253 0.5168 0.8253 0.9084
No log 0.9577 68 0.8016 0.4433 0.8016 0.8953
No log 0.9859 70 0.7653 0.4554 0.7653 0.8748
No log 1.0141 72 0.7460 0.5302 0.7460 0.8637
No log 1.0423 74 0.8003 0.4472 0.8003 0.8946
No log 1.0704 76 0.9858 0.3460 0.9858 0.9929
No log 1.0986 78 0.8542 0.4794 0.8542 0.9242
No log 1.1268 80 0.7238 0.5763 0.7238 0.8508
No log 1.1549 82 0.8300 0.5411 0.8300 0.9110
No log 1.1831 84 1.1047 0.4227 1.1047 1.0510
No log 1.2113 86 0.9343 0.4377 0.9343 0.9666
No log 1.2394 88 0.7116 0.5759 0.7116 0.8436
No log 1.2676 90 0.7813 0.5098 0.7813 0.8839
No log 1.2958 92 0.7672 0.5329 0.7672 0.8759
No log 1.3239 94 0.7380 0.5316 0.7380 0.8591
No log 1.3521 96 0.8070 0.5134 0.8070 0.8983
No log 1.3803 98 0.9355 0.4458 0.9355 0.9672
No log 1.4085 100 0.8304 0.5143 0.8304 0.9112
No log 1.4366 102 0.7099 0.4762 0.7099 0.8425
No log 1.4648 104 0.7278 0.5199 0.7278 0.8531
No log 1.4930 106 0.6908 0.5171 0.6908 0.8312
No log 1.5211 108 0.6845 0.5937 0.6845 0.8274
No log 1.5493 110 0.8151 0.4877 0.8151 0.9029
No log 1.5775 112 0.7328 0.5147 0.7328 0.8560
No log 1.6056 114 0.6658 0.5536 0.6658 0.8160
No log 1.6338 116 0.9179 0.4864 0.9179 0.9581
No log 1.6620 118 0.9262 0.4363 0.9262 0.9624
No log 1.6901 120 0.7520 0.4624 0.7520 0.8672
No log 1.7183 122 0.6901 0.4835 0.6901 0.8307
No log 1.7465 124 0.8176 0.4888 0.8176 0.9042
No log 1.7746 126 0.8175 0.4888 0.8175 0.9042
No log 1.8028 128 0.7062 0.5455 0.7062 0.8404
No log 1.8310 130 0.7284 0.4953 0.7284 0.8534
No log 1.8592 132 0.7719 0.4656 0.7719 0.8786
No log 1.8873 134 0.6733 0.5316 0.6733 0.8206
No log 1.9155 136 0.7307 0.5651 0.7307 0.8548
No log 1.9437 138 1.0072 0.4123 1.0072 1.0036
No log 1.9718 140 1.0129 0.4111 1.0129 1.0064
No log 2.0 142 0.8363 0.5252 0.8363 0.9145
No log 2.0282 144 0.7994 0.5176 0.7994 0.8941
No log 2.0563 146 0.7601 0.5113 0.7601 0.8718
No log 2.0845 148 0.7508 0.5361 0.7508 0.8665
No log 2.1127 150 0.7595 0.4468 0.7595 0.8715
No log 2.1408 152 0.7982 0.4291 0.7982 0.8934
No log 2.1690 154 0.8752 0.3273 0.8752 0.9355
No log 2.1972 156 0.9432 0.2424 0.9432 0.9712
No log 2.2254 158 0.9317 0.2424 0.9317 0.9653
No log 2.2535 160 0.8851 0.3862 0.8851 0.9408
No log 2.2817 162 0.8453 0.3963 0.8453 0.9194
No log 2.3099 164 0.8640 0.2376 0.8640 0.9295
No log 2.3380 166 0.8186 0.3265 0.8186 0.9048
No log 2.3662 168 0.7472 0.3915 0.7472 0.8644
No log 2.3944 170 0.7382 0.5522 0.7382 0.8592
No log 2.4225 172 0.8970 0.4902 0.8970 0.9471
No log 2.4507 174 0.9119 0.4716 0.9119 0.9550
No log 2.4789 176 0.7797 0.5840 0.7797 0.8830
No log 2.5070 178 0.6733 0.6431 0.6733 0.8206
No log 2.5352 180 0.7048 0.5961 0.7048 0.8395
No log 2.5634 182 0.6705 0.5554 0.6705 0.8189
No log 2.5915 184 0.6558 0.5882 0.6558 0.8098
No log 2.6197 186 0.6739 0.5879 0.6739 0.8209
No log 2.6479 188 0.7439 0.5455 0.7439 0.8625
No log 2.6761 190 0.7375 0.6045 0.7375 0.8588
No log 2.7042 192 0.6949 0.6244 0.6949 0.8336
No log 2.7324 194 0.6660 0.6125 0.6660 0.8161
No log 2.7606 196 0.7005 0.6299 0.7005 0.8370
No log 2.7887 198 0.8991 0.5425 0.8991 0.9482
No log 2.8169 200 0.9147 0.5076 0.9147 0.9564
No log 2.8451 202 0.7080 0.6194 0.7080 0.8414
No log 2.8732 204 0.6988 0.6296 0.6988 0.8359
No log 2.9014 206 0.8115 0.4587 0.8115 0.9008
No log 2.9296 208 0.7779 0.5561 0.7779 0.8820
No log 2.9577 210 0.7212 0.5192 0.7212 0.8493
No log 2.9859 212 0.7315 0.5546 0.7315 0.8553
No log 3.0141 214 0.7385 0.5419 0.7385 0.8593
No log 3.0423 216 0.7226 0.5319 0.7226 0.8500
No log 3.0704 218 0.7143 0.5206 0.7143 0.8451
No log 3.0986 220 0.7230 0.5787 0.7230 0.8503
No log 3.1268 222 0.8187 0.5484 0.8187 0.9048
No log 3.1549 224 0.8184 0.5521 0.8184 0.9046
No log 3.1831 226 0.7742 0.5292 0.7742 0.8799
No log 3.2113 228 0.7494 0.5202 0.7494 0.8657
No log 3.2394 230 0.7376 0.5202 0.7376 0.8588
No log 3.2676 232 0.6998 0.4589 0.6998 0.8365
No log 3.2958 234 0.6815 0.4658 0.6815 0.8256
No log 3.3239 236 0.6942 0.4467 0.6942 0.8332
No log 3.3521 238 0.6710 0.4813 0.6710 0.8191
No log 3.3803 240 0.6515 0.5422 0.6515 0.8072
No log 3.4085 242 0.6386 0.5960 0.6386 0.7991
No log 3.4366 244 0.6481 0.5314 0.6481 0.8050
No log 3.4648 246 0.7611 0.4595 0.7611 0.8724
No log 3.4930 248 0.8857 0.4476 0.8857 0.9411
No log 3.5211 250 0.7480 0.4916 0.7480 0.8649
No log 3.5493 252 0.6458 0.6636 0.6458 0.8036
No log 3.5775 254 0.6430 0.5582 0.6430 0.8019
No log 3.6056 256 0.6408 0.5845 0.6408 0.8005
No log 3.6338 258 0.6413 0.5868 0.6413 0.8008
No log 3.6620 260 0.6271 0.6046 0.6271 0.7919
No log 3.6901 262 0.6328 0.6133 0.6328 0.7955
No log 3.7183 264 0.6577 0.6092 0.6577 0.8110
No log 3.7465 266 0.6300 0.6054 0.6300 0.7938
No log 3.7746 268 0.6444 0.5548 0.6444 0.8028
No log 3.8028 270 0.6823 0.5005 0.6823 0.8260
No log 3.8310 272 0.6784 0.4884 0.6784 0.8237
No log 3.8592 274 0.6272 0.5548 0.6272 0.7920
No log 3.8873 276 0.6256 0.5961 0.6256 0.7909
No log 3.9155 278 0.6766 0.5005 0.6766 0.8225
No log 3.9437 280 0.6960 0.4780 0.6960 0.8343
No log 3.9718 282 0.6842 0.4402 0.6842 0.8271
No log 4.0 284 0.6738 0.5063 0.6738 0.8209
No log 4.0282 286 0.6807 0.4813 0.6807 0.8250
No log 4.0563 288 0.6942 0.4624 0.6942 0.8332
No log 4.0845 290 0.6687 0.4976 0.6687 0.8178
No log 4.1127 292 0.6744 0.5080 0.6744 0.8212
No log 4.1408 294 0.6716 0.6370 0.6716 0.8195
No log 4.1690 296 0.7008 0.6122 0.7008 0.8371
No log 4.1972 298 0.6818 0.6508 0.6818 0.8257
No log 4.2254 300 0.6809 0.6311 0.6809 0.8252
No log 4.2535 302 0.6729 0.5419 0.6729 0.8203
No log 4.2817 304 0.6910 0.4864 0.6910 0.8313
No log 4.3099 306 0.6666 0.5315 0.6666 0.8165
No log 4.3380 308 0.6824 0.5363 0.6824 0.8261
No log 4.3662 310 0.7410 0.4439 0.7410 0.8608
No log 4.3944 312 0.7189 0.3583 0.7189 0.8479
No log 4.4225 314 0.6871 0.4783 0.6871 0.8289
No log 4.4507 316 0.7060 0.4868 0.7060 0.8403
No log 4.4789 318 0.6923 0.4868 0.6923 0.8321
No log 4.5070 320 0.6577 0.5035 0.6577 0.8110
No log 4.5352 322 0.6673 0.5394 0.6673 0.8169
No log 4.5634 324 0.6713 0.5273 0.6713 0.8193
No log 4.5915 326 0.7247 0.4576 0.7247 0.8513
No log 4.6197 328 0.7664 0.4301 0.7664 0.8755
No log 4.6479 330 0.7363 0.4576 0.7363 0.8581
No log 4.6761 332 0.6977 0.4953 0.6977 0.8353
No log 4.7042 334 0.7264 0.4879 0.7264 0.8523
No log 4.7324 336 0.7194 0.4675 0.7194 0.8481
No log 4.7606 338 0.7346 0.3733 0.7346 0.8571
No log 4.7887 340 0.7976 0.4301 0.7976 0.8931
No log 4.8169 342 0.8769 0.4549 0.8769 0.9364
No log 4.8451 344 0.8208 0.4928 0.8208 0.9060
No log 4.8732 346 0.7133 0.4576 0.7133 0.8446
No log 4.9014 348 0.6957 0.4919 0.6957 0.8341
No log 4.9296 350 0.7072 0.4160 0.7072 0.8409
No log 4.9577 352 0.7602 0.3883 0.7602 0.8719
No log 4.9859 354 0.7789 0.3883 0.7789 0.8825
No log 5.0141 356 0.7852 0.3590 0.7852 0.8861
No log 5.0423 358 0.8172 0.4162 0.8172 0.9040
No log 5.0704 360 0.8163 0.4162 0.8163 0.9035
No log 5.0986 362 0.8065 0.4697 0.8065 0.8981
No log 5.1268 364 0.7916 0.5183 0.7916 0.8897
No log 5.1549 366 0.7039 0.6164 0.7039 0.8390
No log 5.1831 368 0.7004 0.5950 0.7004 0.8369
No log 5.2113 370 0.7305 0.4743 0.7305 0.8547
No log 5.2394 372 0.8546 0.4792 0.8546 0.9245
No log 5.2676 374 0.9651 0.4302 0.9651 0.9824
No log 5.2958 376 0.9072 0.3483 0.9072 0.9525
No log 5.3239 378 0.7863 0.3583 0.7863 0.8867
No log 5.3521 380 0.7553 0.4384 0.7553 0.8691
No log 5.3803 382 0.7672 0.5357 0.7672 0.8759
No log 5.4085 384 0.7329 0.4640 0.7329 0.8561
No log 5.4366 386 0.7013 0.4706 0.7013 0.8374
No log 5.4648 388 0.7763 0.6240 0.7763 0.8811
No log 5.4930 390 0.8514 0.5343 0.8514 0.9227
No log 5.5211 392 0.7728 0.5895 0.7728 0.8791
No log 5.5493 394 0.7053 0.5969 0.7053 0.8398
No log 5.5775 396 0.7427 0.5128 0.7427 0.8618
No log 5.6056 398 0.8799 0.4729 0.8799 0.9380
No log 5.6338 400 0.8015 0.4820 0.8015 0.8952
No log 5.6620 402 0.6827 0.4498 0.6827 0.8262
No log 5.6901 404 0.7112 0.4485 0.7112 0.8433
No log 5.7183 406 0.7908 0.5266 0.7908 0.8893
No log 5.7465 408 0.7822 0.5387 0.7822 0.8844
No log 5.7746 410 0.7758 0.5387 0.7758 0.8808
No log 5.8028 412 0.7669 0.5147 0.7669 0.8757
No log 5.8310 414 0.7967 0.4926 0.7967 0.8926
No log 5.8592 416 0.7891 0.4804 0.7891 0.8883
No log 5.8873 418 0.7316 0.5673 0.7316 0.8553
No log 5.9155 420 0.7055 0.5673 0.7055 0.8399
No log 5.9437 422 0.6602 0.6165 0.6602 0.8126
No log 5.9718 424 0.6567 0.5735 0.6567 0.8104
No log 6.0 426 0.6677 0.5835 0.6677 0.8171
No log 6.0282 428 0.6946 0.5850 0.6946 0.8334
No log 6.0563 430 0.7458 0.4439 0.7458 0.8636
No log 6.0845 432 0.7314 0.4439 0.7314 0.8552
No log 6.1127 434 0.6785 0.5394 0.6785 0.8237
No log 6.1408 436 0.6763 0.5165 0.6763 0.8224
No log 6.1690 438 0.6723 0.4909 0.6723 0.8200
No log 6.1972 440 0.6809 0.4783 0.6809 0.8252
No log 6.2254 442 0.6863 0.4783 0.6863 0.8284
No log 6.2535 444 0.6917 0.4658 0.6917 0.8317
No log 6.2817 446 0.6798 0.4658 0.6798 0.8245
No log 6.3099 448 0.6504 0.5057 0.6504 0.8065
No log 6.3380 450 0.6309 0.5057 0.6309 0.7943
No log 6.3662 452 0.6185 0.6076 0.6185 0.7865
No log 6.3944 454 0.6142 0.6164 0.6142 0.7837
No log 6.4225 456 0.6162 0.5950 0.6162 0.7850
No log 6.4507 458 0.6206 0.5548 0.6206 0.7878
No log 6.4789 460 0.6216 0.5316 0.6216 0.7884
No log 6.5070 462 0.6544 0.4851 0.6544 0.8089
No log 6.5352 464 0.7355 0.4812 0.7355 0.8576
No log 6.5634 466 0.7984 0.5164 0.7984 0.8936
No log 6.5915 468 0.7782 0.4815 0.7782 0.8822
No log 6.6197 470 0.7281 0.4715 0.7281 0.8533
No log 6.6479 472 0.6765 0.4475 0.6765 0.8225
No log 6.6761 474 0.6795 0.5088 0.6795 0.8243
No log 6.7042 476 0.7505 0.4797 0.7505 0.8663
No log 6.7324 478 0.7352 0.4797 0.7352 0.8575
No log 6.7606 480 0.7041 0.5243 0.7041 0.8391
No log 6.7887 482 0.6590 0.4767 0.6590 0.8118
No log 6.8169 484 0.6777 0.4180 0.6777 0.8232
No log 6.8451 486 0.7070 0.4300 0.7070 0.8408
No log 6.8732 488 0.6979 0.4031 0.6979 0.8354
No log 6.9014 490 0.6872 0.4767 0.6872 0.8290
No log 6.9296 492 0.6973 0.4797 0.6973 0.8350
No log 6.9577 494 0.6901 0.4797 0.6901 0.8308
No log 6.9859 496 0.6737 0.4797 0.6737 0.8208
No log 7.0141 498 0.6585 0.4909 0.6585 0.8115
0.294 7.0423 500 0.6778 0.5218 0.6778 0.8233
0.294 7.0704 502 0.6887 0.5206 0.6887 0.8299
0.294 7.0986 504 0.6671 0.5206 0.6671 0.8168
0.294 7.1268 506 0.6587 0.5605 0.6587 0.8116
0.294 7.1549 508 0.6667 0.5495 0.6667 0.8165
0.294 7.1831 510 0.6767 0.5123 0.6767 0.8226
0.294 7.2113 512 0.6797 0.4871 0.6797 0.8245
0.294 7.2394 514 0.6744 0.4622 0.6744 0.8212
0.294 7.2676 516 0.6826 0.4622 0.6826 0.8262
0.294 7.2958 518 0.6808 0.4622 0.6808 0.8251

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_B_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k14_task5_organization

Finetuned
(4019)
this model