ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k10_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.7886
  • Qwk: -0.0160
  • Mse: 0.7886
  • Rmse: 0.8880

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0385 2 3.8755 0.0017 3.8755 1.9686
No log 0.0769 4 2.3778 0.0213 2.3778 1.5420
No log 0.1154 6 1.3323 0.0 1.3323 1.1542
No log 0.1538 8 1.0532 -0.0468 1.0532 1.0263
No log 0.1923 10 1.0578 -0.0234 1.0578 1.0285
No log 0.2308 12 0.7886 -0.0390 0.7886 0.8880
No log 0.2692 14 0.6581 0.0506 0.6581 0.8112
No log 0.3077 16 0.7383 0.1097 0.7383 0.8593
No log 0.3462 18 0.9479 0.0353 0.9479 0.9736
No log 0.3846 20 1.5216 0.0446 1.5216 1.2335
No log 0.4231 22 1.0908 0.0651 1.0908 1.0444
No log 0.4615 24 0.7305 0.0 0.7305 0.8547
No log 0.5 26 0.8088 0.0 0.8088 0.8993
No log 0.5385 28 0.7614 0.0506 0.7614 0.8726
No log 0.5769 30 0.8265 -0.1241 0.8265 0.9091
No log 0.6154 32 0.9450 -0.1261 0.9450 0.9721
No log 0.6538 34 0.8142 0.0225 0.8142 0.9024
No log 0.6923 36 0.7332 0.0416 0.7332 0.8563
No log 0.7308 38 0.7507 0.0374 0.7507 0.8664
No log 0.7692 40 0.7786 0.0460 0.7786 0.8824
No log 0.8077 42 1.0345 0.0847 1.0345 1.0171
No log 0.8462 44 1.2143 0.0256 1.2143 1.1020
No log 0.8846 46 0.8376 0.0437 0.8376 0.9152
No log 0.9231 48 0.8719 0.0392 0.8719 0.9337
No log 0.9615 50 0.8065 0.0474 0.8065 0.8980
No log 1.0 52 0.9191 0.0727 0.9191 0.9587
No log 1.0385 54 1.4938 0.0799 1.4938 1.2222
No log 1.0769 56 1.0795 -0.0476 1.0795 1.0390
No log 1.1154 58 1.0314 0.0435 1.0314 1.0156
No log 1.1538 60 1.1646 0.0566 1.1646 1.0792
No log 1.1923 62 1.0814 0.0145 1.0814 1.0399
No log 1.2308 64 0.9732 0.1169 0.9732 0.9865
No log 1.2692 66 1.1047 0.0288 1.1047 1.0510
No log 1.3077 68 0.9790 0.1597 0.9790 0.9895
No log 1.3462 70 1.0361 0.0979 1.0361 1.0179
No log 1.3846 72 1.0889 0.0501 1.0889 1.0435
No log 1.4231 74 1.0627 0.0752 1.0627 1.0309
No log 1.4615 76 0.8568 0.2239 0.8568 0.9256
No log 1.5 78 0.9232 0.0424 0.9232 0.9608
No log 1.5385 80 0.8695 0.0697 0.8695 0.9324
No log 1.5769 82 1.0753 0.0081 1.0753 1.0370
No log 1.6154 84 1.0113 0.0875 1.0113 1.0056
No log 1.6538 86 0.9838 0.0856 0.9838 0.9919
No log 1.6923 88 1.0110 0.0792 1.0110 1.0055
No log 1.7308 90 1.1087 0.0319 1.1087 1.0529
No log 1.7692 92 1.3371 0.0519 1.3371 1.1563
No log 1.8077 94 1.0424 0.0366 1.0424 1.0210
No log 1.8462 96 1.0746 -0.0157 1.0746 1.0366
No log 1.8846 98 1.2130 0.0342 1.2130 1.1014
No log 1.9231 100 0.8183 -0.0316 0.8183 0.9046
No log 1.9615 102 0.9823 0.0147 0.9823 0.9911
No log 2.0 104 1.2866 0.0089 1.2866 1.1343
No log 2.0385 106 1.1661 0.0855 1.1661 1.0799
No log 2.0769 108 0.8184 0.0410 0.8184 0.9046
No log 2.1154 110 0.8319 0.0152 0.8319 0.9121
No log 2.1538 112 0.8634 -0.0045 0.8634 0.9292
No log 2.1923 114 1.0005 0.1008 1.0005 1.0002
No log 2.2308 116 0.9724 0.0832 0.9724 0.9861
No log 2.2692 118 1.0235 0.0982 1.0235 1.0117
No log 2.3077 120 1.2043 0.0227 1.2043 1.0974
No log 2.3462 122 1.1508 0.0184 1.1508 1.0728
No log 2.3846 124 1.0158 0.0988 1.0158 1.0078
No log 2.4231 126 1.0522 0.1196 1.0522 1.0258
No log 2.4615 128 1.1568 0.0458 1.1568 1.0756
No log 2.5 130 1.0613 0.0437 1.0613 1.0302
No log 2.5385 132 0.8659 -0.0573 0.8659 0.9305
No log 2.5769 134 0.8316 -0.0331 0.8316 0.9119
No log 2.6154 136 0.7766 -0.0695 0.7766 0.8813
No log 2.6538 138 0.7989 -0.1463 0.7989 0.8938
No log 2.6923 140 0.8170 -0.2306 0.8170 0.9039
No log 2.7308 142 0.8647 -0.0985 0.8647 0.9299
No log 2.7692 144 0.9722 0.0532 0.9722 0.9860
No log 2.8077 146 1.1359 0.0445 1.1359 1.0658
No log 2.8462 148 1.1924 0.0786 1.1924 1.0920
No log 2.8846 150 1.2409 0.0786 1.2409 1.1140
No log 2.9231 152 1.1995 0.0786 1.1995 1.0952
No log 2.9615 154 1.0392 -0.0210 1.0392 1.0194
No log 3.0 156 0.8904 -0.0073 0.8904 0.9436
No log 3.0385 158 0.8023 -0.0658 0.8023 0.8957
No log 3.0769 160 0.8326 0.0214 0.8326 0.9124
No log 3.1154 162 0.8204 0.0214 0.8204 0.9057
No log 3.1538 164 0.9231 0.0540 0.9231 0.9608
No log 3.1923 166 1.0858 0.1469 1.0858 1.0420
No log 3.2308 168 1.1021 0.1820 1.1021 1.0498
No log 3.2692 170 1.1043 0.1408 1.1043 1.0508
No log 3.3077 172 1.1924 0.12 1.1924 1.0919
No log 3.3462 174 1.3183 0.0804 1.3183 1.1482
No log 3.3846 176 1.0937 0.1492 1.0937 1.0458
No log 3.4231 178 0.8857 0.0087 0.8857 0.9411
No log 3.4615 180 0.8331 0.0214 0.8331 0.9127
No log 3.5 182 0.8055 0.0214 0.8055 0.8975
No log 3.5385 184 0.8473 -0.0228 0.8473 0.9205
No log 3.5769 186 0.8825 0.0239 0.8825 0.9394
No log 3.6154 188 0.8361 -0.0054 0.8361 0.9144
No log 3.6538 190 0.8726 0.0628 0.8726 0.9341
No log 3.6923 192 0.8850 0.0214 0.8850 0.9408
No log 3.7308 194 0.9398 0.0569 0.9398 0.9694
No log 3.7692 196 1.1354 -0.0058 1.1354 1.0655
No log 3.8077 198 1.1576 -0.0075 1.1576 1.0759
No log 3.8462 200 1.0463 -0.0263 1.0463 1.0229
No log 3.8846 202 1.2076 0.1405 1.2076 1.0989
No log 3.9231 204 1.0500 0.0039 1.0500 1.0247
No log 3.9615 206 1.1611 -0.0104 1.1611 1.0776
No log 4.0 208 1.1727 -0.0318 1.1727 1.0829
No log 4.0385 210 1.0715 0.0098 1.0715 1.0352
No log 4.0769 212 0.8922 0.0 0.8922 0.9446
No log 4.1154 214 0.8587 0.0282 0.8587 0.9267
No log 4.1538 216 0.8312 -0.0583 0.8312 0.9117
No log 4.1923 218 0.8032 0.0759 0.8032 0.8962
No log 4.2308 220 0.8236 -0.0026 0.8236 0.9075
No log 4.2692 222 0.8816 0.0537 0.8816 0.9389
No log 4.3077 224 0.9062 0.0694 0.9062 0.9519
No log 4.3462 226 0.9799 0.0015 0.9799 0.9899
No log 4.3846 228 0.9236 0.0632 0.9236 0.9611
No log 4.4231 230 1.0717 0.0746 1.0717 1.0352
No log 4.4615 232 1.1037 0.0219 1.1037 1.0506
No log 4.5 234 0.8619 -0.0133 0.8619 0.9284
No log 4.5385 236 0.7646 -0.0240 0.7646 0.8744
No log 4.5769 238 0.9384 0.1316 0.9384 0.9687
No log 4.6154 240 0.9891 0.1152 0.9891 0.9945
No log 4.6538 242 0.7982 0.0628 0.7982 0.8934
No log 4.6923 244 0.8645 0.1078 0.8645 0.9298
No log 4.7308 246 0.8668 0.1078 0.8668 0.9310
No log 4.7692 248 0.8250 -0.1077 0.8250 0.9083
No log 4.8077 250 0.8645 -0.1083 0.8645 0.9298
No log 4.8462 252 0.8403 -0.0127 0.8403 0.9167
No log 4.8846 254 0.8631 0.0119 0.8631 0.9291
No log 4.9231 256 0.8271 0.0327 0.8271 0.9095
No log 4.9615 258 0.8021 0.0 0.8021 0.8956
No log 5.0 260 0.7580 0.0394 0.7580 0.8706
No log 5.0385 262 0.7499 -0.0513 0.7499 0.8660
No log 5.0769 264 0.7369 -0.0030 0.7369 0.8584
No log 5.1154 266 0.7342 0.1254 0.7342 0.8569
No log 5.1538 268 0.7523 0.0680 0.7523 0.8674
No log 5.1923 270 0.7766 0.0236 0.7766 0.8813
No log 5.2308 272 0.8833 0.0999 0.8833 0.9398
No log 5.2692 274 1.0293 0.0723 1.0293 1.0145
No log 5.3077 276 0.9713 0.0717 0.9713 0.9855
No log 5.3462 278 0.8309 -0.0614 0.8309 0.9115
No log 5.3846 280 0.8079 -0.0567 0.8079 0.8989
No log 5.4231 282 0.8951 0.0714 0.8951 0.9461
No log 5.4615 284 1.2189 0.0305 1.2189 1.1040
No log 5.5 286 1.1407 0.0574 1.1407 1.0680
No log 5.5385 288 0.8792 0.0268 0.8792 0.9377
No log 5.5769 290 0.8440 -0.1100 0.8440 0.9187
No log 5.6154 292 0.8824 0.0172 0.8824 0.9394
No log 5.6538 294 0.9866 0.0721 0.9866 0.9933
No log 5.6923 296 0.9587 0.0714 0.9587 0.9791
No log 5.7308 298 0.8480 -0.0280 0.8480 0.9208
No log 5.7692 300 0.8364 -0.0138 0.8364 0.9145
No log 5.8077 302 0.8660 0.0600 0.8660 0.9306
No log 5.8462 304 0.8225 0.0205 0.8225 0.9069
No log 5.8846 306 0.8145 -0.0488 0.8145 0.9025
No log 5.9231 308 0.8096 -0.0614 0.8096 0.8998
No log 5.9615 310 0.7918 -0.0532 0.7918 0.8898
No log 6.0 312 0.7870 -0.1047 0.7870 0.8871
No log 6.0385 314 0.8148 -0.0967 0.8148 0.9027
No log 6.0769 316 0.8335 0.0236 0.8335 0.9130
No log 6.1154 318 0.9269 0.0062 0.9269 0.9628
No log 6.1538 320 0.9146 -0.0192 0.9146 0.9563
No log 6.1923 322 1.0053 -0.0269 1.0053 1.0026
No log 6.2308 324 1.0637 0.0462 1.0637 1.0313
No log 6.2692 326 0.8801 -0.0996 0.8801 0.9381
No log 6.3077 328 0.8066 0.0670 0.8066 0.8981
No log 6.3462 330 0.8002 0.0549 0.8002 0.8945
No log 6.3846 332 0.7659 0.0814 0.7659 0.8752
No log 6.4231 334 0.7826 -0.0086 0.7826 0.8846
No log 6.4615 336 0.8564 0.0 0.8564 0.9254
No log 6.5 338 1.0476 0.0378 1.0476 1.0235
No log 6.5385 340 1.1695 0.0121 1.1695 1.0814
No log 6.5769 342 1.1341 0.0104 1.1341 1.0649
No log 6.6154 344 0.9461 0.0603 0.9461 0.9727
No log 6.6538 346 0.9305 0.0421 0.9305 0.9646
No log 6.6923 348 1.0378 -0.0169 1.0378 1.0187
No log 6.7308 350 0.8836 0.0476 0.8836 0.9400
No log 6.7692 352 0.7596 0.0375 0.7596 0.8716
No log 6.8077 354 0.8364 -0.0226 0.8364 0.9146
No log 6.8462 356 0.8425 -0.0195 0.8425 0.9179
No log 6.8846 358 0.7807 -0.0921 0.7807 0.8836
No log 6.9231 360 0.8104 0.0600 0.8104 0.9002
No log 6.9615 362 0.8278 0.1001 0.8278 0.9098
No log 7.0 364 0.8375 0.0051 0.8375 0.9152
No log 7.0385 366 0.9185 0.0296 0.9185 0.9584
No log 7.0769 368 0.8974 0.0265 0.8974 0.9473
No log 7.1154 370 0.8209 -0.0444 0.8209 0.9060
No log 7.1538 372 0.7939 -0.0113 0.7939 0.8910
No log 7.1923 374 0.7987 -0.0879 0.7987 0.8937
No log 7.2308 376 0.8146 -0.0879 0.8146 0.9026
No log 7.2692 378 0.8267 -0.0163 0.8267 0.9092
No log 7.3077 380 0.8552 0.0660 0.8552 0.9248
No log 7.3462 382 0.8828 -0.0978 0.8828 0.9396
No log 7.3846 384 0.8976 -0.0978 0.8976 0.9474
No log 7.4231 386 0.9402 -0.0630 0.9402 0.9696
No log 7.4615 388 0.8995 -0.0200 0.8995 0.9484
No log 7.5 390 0.8385 -0.0898 0.8385 0.9157
No log 7.5385 392 0.8521 -0.0661 0.8521 0.9231
No log 7.5769 394 0.8034 -0.1813 0.8034 0.8963
No log 7.6154 396 0.7719 0.0759 0.7719 0.8786
No log 7.6538 398 0.8284 0.0191 0.8284 0.9102
No log 7.6923 400 0.8082 0.0225 0.8082 0.8990
No log 7.7308 402 0.8534 -0.0730 0.8534 0.9238
No log 7.7692 404 0.8716 0.0172 0.8716 0.9336
No log 7.8077 406 0.8182 -0.1463 0.8182 0.9046
No log 7.8462 408 0.8240 -0.1066 0.8240 0.9077
No log 7.8846 410 0.8275 0.0214 0.8275 0.9097
No log 7.9231 412 0.8182 0.0214 0.8182 0.9045
No log 7.9615 414 0.7984 -0.0170 0.7984 0.8935
No log 8.0 416 0.8279 -0.1332 0.8279 0.9099
No log 8.0385 418 0.8034 -0.1463 0.8034 0.8963
No log 8.0769 420 0.7597 -0.0152 0.7597 0.8716
No log 8.1154 422 0.8699 0.0909 0.8699 0.9327
No log 8.1538 424 0.9273 0.0755 0.9273 0.9630
No log 8.1923 426 0.8477 0.0512 0.8477 0.9207
No log 8.2308 428 0.8055 -0.1461 0.8055 0.8975
No log 8.2692 430 0.9176 -0.0771 0.9176 0.9579
No log 8.3077 432 0.9113 -0.0820 0.9113 0.9546
No log 8.3462 434 0.8949 -0.0508 0.8949 0.9460
No log 8.3846 436 0.8256 -0.0595 0.8256 0.9086
No log 8.4231 438 0.8323 -0.0170 0.8323 0.9123
No log 8.4615 440 0.8196 -0.0170 0.8196 0.9053
No log 8.5 442 0.8285 -0.0118 0.8285 0.9102
No log 8.5385 444 0.8669 -0.1214 0.8669 0.9311
No log 8.5769 446 0.8675 -0.1214 0.8675 0.9314
No log 8.6154 448 0.8458 -0.0831 0.8458 0.9197
No log 8.6538 450 0.8279 0.0214 0.8279 0.9099
No log 8.6923 452 0.9474 0.0333 0.9474 0.9733
No log 8.7308 454 0.9233 0.0755 0.9233 0.9609
No log 8.7692 456 0.8141 -0.0287 0.8141 0.9023
No log 8.8077 458 0.8167 -0.1397 0.8167 0.9037
No log 8.8462 460 0.8312 -0.0738 0.8312 0.9117
No log 8.8846 462 0.7939 -0.1397 0.7939 0.8910
No log 8.9231 464 0.7705 0.0909 0.7705 0.8778
No log 8.9615 466 0.8129 0.1047 0.8129 0.9016
No log 9.0 468 0.8328 0.0600 0.8328 0.9126
No log 9.0385 470 0.8411 0.0 0.8411 0.9171
No log 9.0769 472 0.9233 -0.0365 0.9233 0.9609
No log 9.1154 474 0.9407 -0.0606 0.9407 0.9699
No log 9.1538 476 0.8557 -0.1102 0.8557 0.9251
No log 9.1923 478 0.8163 0.0269 0.8163 0.9035
No log 9.2308 480 0.8931 0.1291 0.8931 0.9450
No log 9.2692 482 0.9105 0.0826 0.9105 0.9542
No log 9.3077 484 0.8572 0.0205 0.8572 0.9258
No log 9.3462 486 0.8553 0.0226 0.8553 0.9248
No log 9.3846 488 0.8644 -0.0796 0.8644 0.9297
No log 9.4231 490 0.8709 -0.0717 0.8709 0.9332
No log 9.4615 492 0.8141 -0.0056 0.8141 0.9023
No log 9.5 494 0.8171 0.1627 0.8171 0.9039
No log 9.5385 496 0.8228 0.1097 0.8228 0.9071
No log 9.5769 498 0.7935 0.1202 0.7935 0.8908
0.321 9.6154 500 0.8003 0.1202 0.8003 0.8946
0.321 9.6538 502 0.8150 -0.0524 0.8150 0.9027
0.321 9.6923 504 0.8966 -0.0166 0.8966 0.9469
0.321 9.7308 506 0.9207 0.0734 0.9207 0.9595
0.321 9.7692 508 0.8497 -0.1331 0.8497 0.9218
0.321 9.8077 510 0.7886 -0.0160 0.7886 0.8880

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k10_task3_organization

Finetuned
(4019)
this model