ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k14_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8644
  • Qwk: -0.0439
  • Mse: 0.8644
  • Rmse: 0.9297

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0278 2 3.6501 -0.0058 3.6501 1.9105
No log 0.0556 4 2.3090 0.0050 2.3090 1.5195
No log 0.0833 6 0.8707 0.0458 0.8707 0.9331
No log 0.1111 8 1.1473 -0.0178 1.1473 1.0711
No log 0.1389 10 0.9930 -0.0117 0.9930 0.9965
No log 0.1667 12 0.8499 0.0711 0.8499 0.9219
No log 0.1944 14 0.9389 -0.0617 0.9389 0.9690
No log 0.2222 16 0.9575 -0.0648 0.9575 0.9785
No log 0.25 18 0.8074 0.0225 0.8074 0.8986
No log 0.2778 20 0.7492 -0.0679 0.7492 0.8656
No log 0.3056 22 0.7508 0.0555 0.7508 0.8665
No log 0.3333 24 0.7760 0.0460 0.7760 0.8809
No log 0.3611 26 0.9794 -0.0122 0.9794 0.9896
No log 0.3889 28 1.0713 -0.0583 1.0713 1.0350
No log 0.4167 30 0.9275 -0.2121 0.9275 0.9631
No log 0.4444 32 0.8920 -0.1682 0.8920 0.9445
No log 0.4722 34 0.8047 0.0506 0.8047 0.8970
No log 0.5 36 0.8100 -0.0949 0.8100 0.9000
No log 0.5278 38 0.7826 -0.0274 0.7826 0.8847
No log 0.5556 40 1.0641 0.0134 1.0641 1.0316
No log 0.5833 42 0.8785 0.0562 0.8785 0.9373
No log 0.6111 44 0.9418 0.0964 0.9418 0.9705
No log 0.6389 46 1.0279 0.0326 1.0279 1.0138
No log 0.6667 48 0.8400 0.0690 0.8400 0.9165
No log 0.6944 50 0.8837 -0.0861 0.8837 0.9401
No log 0.7222 52 0.8370 -0.0373 0.8370 0.9149
No log 0.75 54 0.8078 0.1635 0.8078 0.8988
No log 0.7778 56 1.0176 0.1312 1.0176 1.0088
No log 0.8056 58 0.9579 0.1192 0.9579 0.9787
No log 0.8333 60 0.8464 0.1417 0.8464 0.9200
No log 0.8611 62 0.8789 -0.0215 0.8789 0.9375
No log 0.8889 64 0.8079 0.1424 0.8079 0.8988
No log 0.9167 66 0.7611 0.1485 0.7611 0.8724
No log 0.9444 68 0.7174 0.1705 0.7174 0.8470
No log 0.9722 70 0.7329 0.1148 0.7329 0.8561
No log 1.0 72 0.8614 -0.0490 0.8614 0.9281
No log 1.0278 74 0.7964 0.0476 0.7964 0.8924
No log 1.0556 76 0.7970 0.2934 0.7970 0.8928
No log 1.0833 78 0.8844 0.2053 0.8844 0.9404
No log 1.1111 80 1.1242 0.0769 1.1242 1.0603
No log 1.1389 82 1.0367 0.1242 1.0367 1.0182
No log 1.1667 84 1.0239 0.0945 1.0239 1.0119
No log 1.1944 86 1.0741 0.1565 1.0741 1.0364
No log 1.2222 88 0.8950 0.0975 0.8950 0.9461
No log 1.25 90 0.9570 0.1465 0.9570 0.9783
No log 1.2778 92 1.2541 0.1256 1.2541 1.1199
No log 1.3056 94 1.2615 0.1226 1.2615 1.1232
No log 1.3333 96 1.0054 0.1243 1.0054 1.0027
No log 1.3611 98 1.3271 -0.0272 1.3271 1.1520
No log 1.3889 100 1.2912 -0.0323 1.2912 1.1363
No log 1.4167 102 0.9196 0.0509 0.9196 0.9590
No log 1.4444 104 0.8834 0.1132 0.8834 0.9399
No log 1.4722 106 1.1091 0.0980 1.1091 1.0531
No log 1.5 108 1.2480 0.1020 1.2480 1.1171
No log 1.5278 110 1.0400 0.1220 1.0400 1.0198
No log 1.5556 112 0.9945 -0.0259 0.9945 0.9972
No log 1.5833 114 1.5060 0.1016 1.5060 1.2272
No log 1.6111 116 1.4007 0.0614 1.4007 1.1835
No log 1.6389 118 0.9342 0.0105 0.9342 0.9665
No log 1.6667 120 0.9154 0.0166 0.9154 0.9568
No log 1.6944 122 0.9411 0.0941 0.9411 0.9701
No log 1.7222 124 0.8736 0.0741 0.8736 0.9347
No log 1.75 126 0.8742 0.0717 0.8742 0.9350
No log 1.7778 128 0.9071 0.0223 0.9071 0.9524
No log 1.8056 130 0.9740 0.1252 0.9740 0.9869
No log 1.8333 132 0.9572 0.0478 0.9572 0.9784
No log 1.8611 134 0.8899 0.1432 0.8899 0.9433
No log 1.8889 136 0.9164 0.0084 0.9164 0.9573
No log 1.9167 138 1.1875 0.1274 1.1875 1.0897
No log 1.9444 140 1.1496 0.1621 1.1496 1.0722
No log 1.9722 142 0.8556 0.0301 0.8556 0.9250
No log 2.0 144 0.8926 0.0392 0.8926 0.9448
No log 2.0278 146 0.8190 0.0529 0.8190 0.9050
No log 2.0556 148 0.9148 0.0281 0.9148 0.9565
No log 2.0833 150 0.8915 0.0494 0.8915 0.9442
No log 2.1111 152 0.9282 0.1313 0.9282 0.9634
No log 2.1389 154 0.8975 0.1050 0.8975 0.9474
No log 2.1667 156 0.9644 -0.0353 0.9644 0.9820
No log 2.1944 158 1.0115 -0.0154 1.0115 1.0057
No log 2.2222 160 0.9710 -0.0533 0.9710 0.9854
No log 2.25 162 0.7893 0.0879 0.7893 0.8884
No log 2.2778 164 0.7874 0.0776 0.7874 0.8874
No log 2.3056 166 0.7981 0.0869 0.7981 0.8934
No log 2.3333 168 0.8345 0.0804 0.8345 0.9135
No log 2.3611 170 1.0156 -0.0539 1.0156 1.0078
No log 2.3889 172 0.9577 0.0262 0.9577 0.9786
No log 2.4167 174 0.8459 0.0778 0.8459 0.9197
No log 2.4444 176 0.8140 0.0110 0.8140 0.9022
No log 2.4722 178 0.8686 -0.0138 0.8686 0.9320
No log 2.5 180 0.7897 0.0110 0.7897 0.8887
No log 2.5278 182 0.7988 0.1282 0.7988 0.8938
No log 2.5556 184 0.8281 0.0717 0.8281 0.9100
No log 2.5833 186 0.8610 0.0661 0.8610 0.9279
No log 2.6111 188 1.0695 0.0680 1.0695 1.0341
No log 2.6389 190 1.1134 -0.0214 1.1134 1.0552
No log 2.6667 192 0.8740 0.1010 0.8740 0.9349
No log 2.6944 194 0.8760 0.0484 0.8760 0.9360
No log 2.7222 196 0.8204 0.1983 0.8204 0.9058
No log 2.75 198 0.8112 0.1821 0.8112 0.9007
No log 2.7778 200 0.7623 0.1148 0.7623 0.8731
No log 2.8056 202 0.7357 0.1259 0.7357 0.8577
No log 2.8333 204 0.7620 0.0122 0.7620 0.8729
No log 2.8611 206 0.7956 -0.0837 0.7956 0.8919
No log 2.8889 208 0.7537 0.0061 0.7537 0.8682
No log 2.9167 210 0.7313 0.1828 0.7313 0.8551
No log 2.9444 212 0.7613 0.1644 0.7613 0.8725
No log 2.9722 214 0.8153 0.1630 0.8153 0.9029
No log 3.0 216 0.8736 0.0961 0.8736 0.9347
No log 3.0278 218 0.8830 0.1580 0.8830 0.9397
No log 3.0556 220 1.1357 0.0741 1.1357 1.0657
No log 3.0833 222 1.0524 -0.0113 1.0524 1.0259
No log 3.1111 224 0.8244 0.0051 0.8244 0.9080
No log 3.1389 226 0.8012 0.0410 0.8012 0.8951
No log 3.1667 228 0.7994 0.1573 0.7994 0.8941
No log 3.1944 230 0.8486 0.0679 0.8486 0.9212
No log 3.2222 232 0.9120 -0.0377 0.9120 0.9550
No log 3.25 234 1.0292 0.0627 1.0292 1.0145
No log 3.2778 236 1.4643 0.0849 1.4643 1.2101
No log 3.3056 238 1.7168 0.0626 1.7168 1.3103
No log 3.3333 240 1.3297 0.0839 1.3297 1.1531
No log 3.3611 242 0.8975 0.0093 0.8975 0.9473
No log 3.3889 244 0.8426 0.1095 0.8426 0.9179
No log 3.4167 246 0.7888 0.1143 0.7888 0.8881
No log 3.4444 248 0.8240 0.0532 0.8240 0.9077
No log 3.4722 250 0.9160 0.0107 0.9160 0.9571
No log 3.5 252 0.8270 0.0474 0.8270 0.9094
No log 3.5278 254 0.8413 0.0856 0.8413 0.9172
No log 3.5556 256 0.9490 0.0006 0.9490 0.9742
No log 3.5833 258 1.1484 -0.0327 1.1484 1.0716
No log 3.6111 260 1.0105 -0.0159 1.0105 1.0052
No log 3.6389 262 0.8081 0.0481 0.8081 0.8990
No log 3.6667 264 0.7741 0.1298 0.7741 0.8798
No log 3.6944 266 0.7639 0.1304 0.7639 0.8740
No log 3.7222 268 0.7688 0.0973 0.7688 0.8768
No log 3.75 270 0.8073 -0.0307 0.8073 0.8985
No log 3.7778 272 0.8361 -0.1597 0.8361 0.9144
No log 3.8056 274 0.8900 -0.0268 0.8900 0.9434
No log 3.8333 276 0.9905 0.0589 0.9905 0.9952
No log 3.8611 278 1.0331 0.0277 1.0331 1.0164
No log 3.8889 280 1.0923 -0.0589 1.0923 1.0451
No log 3.9167 282 1.0260 -0.0670 1.0260 1.0129
No log 3.9444 284 0.9425 -0.1002 0.9425 0.9708
No log 3.9722 286 0.9227 -0.0634 0.9227 0.9606
No log 4.0 288 0.9243 0.0172 0.9243 0.9614
No log 4.0278 290 0.9523 -0.0861 0.9523 0.9758
No log 4.0556 292 0.9844 -0.0981 0.9844 0.9922
No log 4.0833 294 0.8966 0.0 0.8966 0.9469
No log 4.1111 296 0.9095 -0.0533 0.9095 0.9537
No log 4.1389 298 1.0673 -0.0827 1.0673 1.0331
No log 4.1667 300 1.4333 0.0588 1.4333 1.1972
No log 4.1944 302 1.3132 0.0027 1.3132 1.1460
No log 4.2222 304 0.9816 -0.0197 0.9816 0.9907
No log 4.25 306 0.9838 0.0113 0.9838 0.9919
No log 4.2778 308 0.8932 0.0660 0.8932 0.9451
No log 4.3056 310 0.8958 -0.0687 0.8958 0.9465
No log 4.3333 312 0.8815 -0.0730 0.8815 0.9389
No log 4.3611 314 0.8899 -0.0238 0.8899 0.9433
No log 4.3889 316 0.8555 0.0359 0.8555 0.9249
No log 4.4167 318 0.8687 0.0376 0.8687 0.9320
No log 4.4444 320 0.8980 0.0172 0.8980 0.9476
No log 4.4722 322 0.9434 -0.0085 0.9434 0.9713
No log 4.5 324 0.8689 -0.0195 0.8689 0.9321
No log 4.5278 326 0.7980 0.0 0.7980 0.8933
No log 4.5556 328 0.7977 -0.0030 0.7977 0.8931
No log 4.5833 330 0.8472 0.0344 0.8472 0.9205
No log 4.6111 332 0.9351 -0.0073 0.9351 0.9670
No log 4.6389 334 1.0831 0.0231 1.0831 1.0407
No log 4.6667 336 1.4394 0.0253 1.4394 1.1998
No log 4.6944 338 1.4158 0.0011 1.4158 1.1899
No log 4.7222 340 1.0451 -0.0545 1.0451 1.0223
No log 4.75 342 0.8737 0.0313 0.8737 0.9347
No log 4.7778 344 0.8767 0.0146 0.8767 0.9363
No log 4.8056 346 0.8402 0.1196 0.8402 0.9166
No log 4.8333 348 0.8503 -0.0370 0.8503 0.9221
No log 4.8611 350 0.8704 -0.0647 0.8704 0.9330
No log 4.8889 352 0.8272 0.0428 0.8272 0.9095
No log 4.9167 354 0.8343 0.1362 0.8343 0.9134
No log 4.9444 356 0.8556 0.1573 0.8556 0.9250
No log 4.9722 358 0.9024 0.1310 0.9024 0.9499
No log 5.0 360 0.9912 0.0025 0.9912 0.9956
No log 5.0278 362 1.0809 -0.0462 1.0809 1.0397
No log 5.0556 364 0.9310 0.0281 0.9310 0.9649
No log 5.0833 366 0.8153 0.1354 0.8153 0.9029
No log 5.1111 368 0.8106 0.1354 0.8106 0.9004
No log 5.1389 370 0.8281 0.1882 0.8281 0.9100
No log 5.1667 372 0.8648 -0.0153 0.8648 0.9299
No log 5.1944 374 0.8275 0.1413 0.8275 0.9097
No log 5.2222 376 0.8565 0.0879 0.8565 0.9255
No log 5.25 378 0.9585 0.0296 0.9585 0.9790
No log 5.2778 380 0.9130 -0.0503 0.9130 0.9555
No log 5.3056 382 0.9197 -0.0123 0.9197 0.9590
No log 5.3333 384 0.9329 -0.0247 0.9329 0.9659
No log 5.3611 386 0.9894 -0.0222 0.9894 0.9947
No log 5.3889 388 1.0478 0.0915 1.0478 1.0236
No log 5.4167 390 0.9832 -0.0825 0.9832 0.9916
No log 5.4444 392 1.0340 -0.0198 1.0340 1.0168
No log 5.4722 394 0.9648 0.0070 0.9648 0.9822
No log 5.5 396 0.8360 -0.0329 0.8360 0.9143
No log 5.5278 398 0.7950 0.0571 0.7950 0.8916
No log 5.5556 400 0.7645 0.0 0.7645 0.8744
No log 5.5833 402 0.7554 -0.0160 0.7554 0.8691
No log 5.6111 404 0.8769 0.0456 0.8769 0.9364
No log 5.6389 406 0.8266 0.0562 0.8266 0.9092
No log 5.6667 408 0.7637 0.0479 0.7637 0.8739
No log 5.6944 410 0.8146 0.0141 0.8146 0.9026
No log 5.7222 412 0.8108 0.1453 0.8108 0.9004
No log 5.75 414 0.8502 0.0725 0.8502 0.9221
No log 5.7778 416 0.9594 0.0015 0.9594 0.9795
No log 5.8056 418 0.9401 0.0081 0.9401 0.9696
No log 5.8333 420 0.9013 0.0748 0.9013 0.9494
No log 5.8611 422 1.0378 0.0134 1.0378 1.0187
No log 5.8889 424 1.0248 -0.0191 1.0248 1.0123
No log 5.9167 426 0.9038 0.0955 0.9038 0.9507
No log 5.9444 428 0.8607 0.1646 0.8607 0.9277
No log 5.9722 430 0.8791 0.1744 0.8791 0.9376
No log 6.0 432 0.8899 0.1744 0.8899 0.9433
No log 6.0278 434 0.9364 0.0239 0.9364 0.9677
No log 6.0556 436 1.0478 -0.1111 1.0478 1.0236
No log 6.0833 438 0.9952 -0.0301 0.9952 0.9976
No log 6.1111 440 0.8845 0.0392 0.8845 0.9405
No log 6.1389 442 0.8503 0.1143 0.8503 0.9221
No log 6.1667 444 0.8351 0.0344 0.8351 0.9138
No log 6.1944 446 0.8691 0.0145 0.8691 0.9323
No log 6.2222 448 0.9041 -0.0904 0.9041 0.9508
No log 6.25 450 0.8811 0.0145 0.8811 0.9387
No log 6.2778 452 0.8538 0.0361 0.8538 0.9240
No log 6.3056 454 0.8840 0.0145 0.8840 0.9402
No log 6.3333 456 0.9161 0.0230 0.9161 0.9572
No log 6.3611 458 0.8642 -0.0024 0.8642 0.9296
No log 6.3889 460 0.8715 -0.0099 0.8715 0.9335
No log 6.4167 462 0.8794 0.0437 0.8794 0.9378
No log 6.4444 464 0.9186 0.0916 0.9186 0.9584
No log 6.4722 466 1.0409 -0.0579 1.0409 1.0203
No log 6.5 468 1.0173 0.0598 1.0173 1.0086
No log 6.5278 470 0.9220 0.0842 0.9220 0.9602
No log 6.5556 472 0.8793 0.0514 0.8793 0.9377
No log 6.5833 474 0.8729 0.0592 0.8729 0.9343
No log 6.6111 476 0.8107 0.0926 0.8107 0.9004
No log 6.6389 478 0.8077 0.0874 0.8077 0.8987
No log 6.6667 480 0.8290 0.0879 0.8290 0.9105
No log 6.6944 482 0.8811 0.0145 0.8811 0.9387
No log 6.7222 484 0.9403 -0.0585 0.9403 0.9697
No log 6.75 486 1.0157 -0.0381 1.0157 1.0078
No log 6.7778 488 0.9306 -0.0585 0.9306 0.9647
No log 6.8056 490 0.8609 0.1518 0.8609 0.9279
No log 6.8333 492 0.8195 0.0884 0.8195 0.9053
No log 6.8611 494 0.8697 -0.1354 0.8697 0.9326
No log 6.8889 496 0.8288 0.0218 0.8288 0.9104
No log 6.9167 498 0.7810 0.0814 0.7810 0.8837
0.3045 6.9444 500 0.8009 0.1691 0.8009 0.8950
0.3045 6.9722 502 0.8002 0.0922 0.8002 0.8945
0.3045 7.0 504 0.9374 0.0337 0.9374 0.9682
0.3045 7.0278 506 1.0163 0.0448 1.0163 1.0081
0.3045 7.0556 508 0.8986 -0.0103 0.8986 0.9479
0.3045 7.0833 510 0.8031 0.1860 0.8031 0.8961
0.3045 7.1111 512 0.7845 0.1146 0.7845 0.8857
0.3045 7.1389 514 0.7856 0.1362 0.7856 0.8863
0.3045 7.1667 516 0.8506 0.0275 0.8506 0.9223
0.3045 7.1944 518 0.8589 -0.0424 0.8589 0.9267
0.3045 7.2222 520 0.8644 -0.0439 0.8644 0.9297

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run3_AugV5_k14_task3_organization

Finetuned
(4019)
this model