ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k2_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9171
  • Qwk: 0.0486
  • Mse: 0.9171
  • Rmse: 0.9577

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.4 2 3.5983 -0.0154 3.5983 1.8969
No log 0.8 4 2.0367 0.0943 2.0367 1.4271
No log 1.2 6 1.6440 0.0168 1.6440 1.2822
No log 1.6 8 1.1548 -0.0149 1.1548 1.0746
No log 2.0 10 0.9063 -0.1267 0.9063 0.9520
No log 2.4 12 0.9291 -0.0571 0.9291 0.9639
No log 2.8 14 1.0005 -0.0301 1.0005 1.0003
No log 3.2 16 1.1130 -0.0686 1.1130 1.0550
No log 3.6 18 1.4395 -0.0500 1.4395 1.1998
No log 4.0 20 1.1993 -0.0423 1.1993 1.0951
No log 4.4 22 0.9036 -0.1268 0.9036 0.9506
No log 4.8 24 0.8924 -0.0474 0.8924 0.9446
No log 5.2 26 0.9055 0.0576 0.9055 0.9516
No log 5.6 28 1.0385 0.0329 1.0385 1.0191
No log 6.0 30 0.8930 0.0377 0.8930 0.9450
No log 6.4 32 0.7882 -0.0264 0.7882 0.8878
No log 6.8 34 1.1315 -0.0143 1.1315 1.0637
No log 7.2 36 2.1501 -0.0109 2.1501 1.4663
No log 7.6 38 1.7336 -0.0155 1.7336 1.3167
No log 8.0 40 1.0045 -0.1628 1.0045 1.0023
No log 8.4 42 0.9302 -0.0930 0.9302 0.9645
No log 8.8 44 1.0225 -0.1710 1.0225 1.0112
No log 9.2 46 1.5410 0.0636 1.5410 1.2414
No log 9.6 48 1.8504 -0.0236 1.8504 1.3603
No log 10.0 50 1.3787 0.0390 1.3787 1.1742
No log 10.4 52 1.0126 -0.0454 1.0126 1.0063
No log 10.8 54 0.9488 -0.2635 0.9488 0.9741
No log 11.2 56 1.0185 -0.0778 1.0185 1.0092
No log 11.6 58 1.3028 -0.0582 1.3028 1.1414
No log 12.0 60 1.8171 0.0251 1.8171 1.3480
No log 12.4 62 1.5871 -0.0615 1.5871 1.2598
No log 12.8 64 1.0228 -0.0008 1.0228 1.0113
No log 13.2 66 0.9067 -0.0583 0.9067 0.9522
No log 13.6 68 0.9537 -0.1140 0.9537 0.9766
No log 14.0 70 1.2030 -0.0518 1.2030 1.0968
No log 14.4 72 1.1492 -0.0518 1.1492 1.0720
No log 14.8 74 0.9236 -0.0316 0.9236 0.9610
No log 15.2 76 0.8700 -0.0672 0.8700 0.9327
No log 15.6 78 0.9809 -0.0031 0.9809 0.9904
No log 16.0 80 1.3016 -0.0049 1.3016 1.1409
No log 16.4 82 1.2637 0.0252 1.2637 1.1241
No log 16.8 84 0.9376 0.0456 0.9376 0.9683
No log 17.2 86 0.8190 -0.0658 0.8190 0.9050
No log 17.6 88 0.8380 -0.0672 0.8380 0.9154
No log 18.0 90 0.9959 0.0016 0.9959 0.9979
No log 18.4 92 1.2001 -0.0261 1.2001 1.0955
No log 18.8 94 1.1236 0.0152 1.1236 1.0600
No log 19.2 96 0.8954 -0.0711 0.8954 0.9463
No log 19.6 98 0.8370 -0.1054 0.8370 0.9149
No log 20.0 100 0.8572 0.0 0.8572 0.9258
No log 20.4 102 0.8652 -0.1026 0.8652 0.9302
No log 20.8 104 1.0694 0.0508 1.0694 1.0341
No log 21.2 106 1.4283 -0.0957 1.4283 1.1951
No log 21.6 108 1.4024 -0.0697 1.4024 1.1842
No log 22.0 110 1.1356 0.0815 1.1356 1.0656
No log 22.4 112 0.9289 -0.0362 0.9289 0.9638
No log 22.8 114 0.8668 0.0118 0.8668 0.9310
No log 23.2 116 0.8774 0.0277 0.8774 0.9367
No log 23.6 118 0.9029 0.0946 0.9029 0.9502
No log 24.0 120 0.8757 0.0392 0.8757 0.9358
No log 24.4 122 0.9018 0.0304 0.9018 0.9496
No log 24.8 124 0.9460 0.0224 0.9460 0.9726
No log 25.2 126 1.0996 0.1458 1.0996 1.0486
No log 25.6 128 1.1325 0.0446 1.1325 1.0642
No log 26.0 130 0.9464 0.1077 0.9464 0.9728
No log 26.4 132 0.8606 -0.0054 0.8606 0.9277
No log 26.8 134 0.8225 0.0040 0.8225 0.9069
No log 27.2 136 0.8432 0.0016 0.8432 0.9183
No log 27.6 138 0.9449 0.0200 0.9449 0.9720
No log 28.0 140 0.9887 0.0747 0.9887 0.9943
No log 28.4 142 0.9456 0.0458 0.9456 0.9724
No log 28.8 144 0.8341 -0.0735 0.8341 0.9133
No log 29.2 146 0.7990 -0.0262 0.7990 0.8939
No log 29.6 148 0.8100 -0.0262 0.8100 0.9000
No log 30.0 150 0.8735 -0.0393 0.8735 0.9346
No log 30.4 152 1.0502 0.0086 1.0502 1.0248
No log 30.8 154 1.1320 0.0881 1.1320 1.0639
No log 31.2 156 1.0608 0.0659 1.0608 1.0299
No log 31.6 158 0.9740 -0.0877 0.9740 0.9869
No log 32.0 160 0.8866 0.0719 0.8866 0.9416
No log 32.4 162 0.8419 0.0065 0.8419 0.9175
No log 32.8 164 0.8100 0.0146 0.8100 0.9000
No log 33.2 166 0.7989 -0.0672 0.7989 0.8938
No log 33.6 168 0.8020 -0.0303 0.8020 0.8955
No log 34.0 170 0.8306 0.0377 0.8306 0.9114
No log 34.4 172 0.8601 -0.1554 0.8601 0.9274
No log 34.8 174 0.8982 -0.1166 0.8982 0.9477
No log 35.2 176 0.9173 -0.1172 0.9173 0.9578
No log 35.6 178 0.9725 0.0431 0.9725 0.9861
No log 36.0 180 1.0283 0.0107 1.0283 1.0140
No log 36.4 182 0.9647 -0.0815 0.9647 0.9822
No log 36.8 184 0.9598 -0.0806 0.9598 0.9797
No log 37.2 186 1.0847 -0.0827 1.0847 1.0415
No log 37.6 188 1.1158 -0.0320 1.1158 1.0563
No log 38.0 190 1.0559 0.0353 1.0559 1.0276
No log 38.4 192 0.9416 -0.1175 0.9416 0.9703
No log 38.8 194 0.8557 -0.0341 0.8557 0.9250
No log 39.2 196 0.8110 0.0269 0.8110 0.9006
No log 39.6 198 0.8142 0.0236 0.8142 0.9023
No log 40.0 200 0.8759 -0.0393 0.8759 0.9359
No log 40.4 202 1.0406 0.1223 1.0406 1.0201
No log 40.8 204 1.2372 -0.0149 1.2372 1.1123
No log 41.2 206 1.2434 -0.0149 1.2434 1.1151
No log 41.6 208 1.0998 0.0493 1.0998 1.0487
No log 42.0 210 0.9365 -0.0454 0.9365 0.9678
No log 42.4 212 0.9010 -0.0376 0.9010 0.9492
No log 42.8 214 0.9302 -0.0787 0.9302 0.9645
No log 43.2 216 0.9733 -0.1172 0.9733 0.9866
No log 43.6 218 1.0557 -0.0187 1.0557 1.0275
No log 44.0 220 1.0370 -0.0855 1.0370 1.0183
No log 44.4 222 0.9934 -0.0508 0.9934 0.9967
No log 44.8 224 0.9851 -0.0892 0.9851 0.9925
No log 45.2 226 0.9691 -0.0218 0.9691 0.9845
No log 45.6 228 0.9413 -0.0870 0.9413 0.9702
No log 46.0 230 0.9093 -0.0815 0.9093 0.9536
No log 46.4 232 0.8848 -0.0008 0.8848 0.9406
No log 46.8 234 0.9149 -0.0424 0.9149 0.9565
No log 47.2 236 1.0035 -0.0832 1.0035 1.0018
No log 47.6 238 0.9920 -0.0468 0.9920 0.9960
No log 48.0 240 0.9809 -0.0468 0.9809 0.9904
No log 48.4 242 0.9711 -0.0440 0.9711 0.9854
No log 48.8 244 0.9566 -0.0052 0.9566 0.9780
No log 49.2 246 1.0320 -0.0521 1.0320 1.0159
No log 49.6 248 1.0535 0.0416 1.0535 1.0264
No log 50.0 250 1.0641 0.0319 1.0641 1.0315
No log 50.4 252 0.9961 0.0569 0.9961 0.9980
No log 50.8 254 0.8866 0.1042 0.8866 0.9416
No log 51.2 256 0.7958 0.1342 0.7958 0.8921
No log 51.6 258 0.7759 -0.0170 0.7759 0.8808
No log 52.0 260 0.7789 -0.0628 0.7789 0.8825
No log 52.4 262 0.8041 -0.0240 0.8041 0.8967
No log 52.8 264 0.8556 0.0719 0.8556 0.9250
No log 53.2 266 0.9291 -0.0200 0.9291 0.9639
No log 53.6 268 0.9661 -0.0245 0.9661 0.9829
No log 54.0 270 0.9705 -0.1230 0.9705 0.9852
No log 54.4 272 0.9377 -0.0454 0.9377 0.9684
No log 54.8 274 0.8969 -0.0376 0.8969 0.9471
No log 55.2 276 0.8771 -0.0359 0.8771 0.9365
No log 55.6 278 0.8681 0.0065 0.8681 0.9317
No log 56.0 280 0.8850 -0.0359 0.8850 0.9407
No log 56.4 282 0.9111 -0.0376 0.9111 0.9545
No log 56.8 284 0.9662 -0.0440 0.9662 0.9830
No log 57.2 286 1.0560 0.0391 1.0560 1.0276
No log 57.6 288 1.1238 0.0543 1.1238 1.0601
No log 58.0 290 1.1666 0.0469 1.1666 1.0801
No log 58.4 292 1.1103 0.0819 1.1103 1.0537
No log 58.8 294 1.0388 0.0282 1.0388 1.0192
No log 59.2 296 0.9472 -0.0218 0.9472 0.9732
No log 59.6 298 0.8609 0.0333 0.8609 0.9278
No log 60.0 300 0.8283 -0.0723 0.8283 0.9101
No log 60.4 302 0.8041 -0.0699 0.8041 0.8967
No log 60.8 304 0.8163 -0.0735 0.8163 0.9035
No log 61.2 306 0.8523 -0.0818 0.8523 0.9232
No log 61.6 308 0.8916 -0.0138 0.8916 0.9443
No log 62.0 310 0.9494 -0.0211 0.9494 0.9744
No log 62.4 312 0.9733 -0.0211 0.9733 0.9865
No log 62.8 314 0.9678 -0.0211 0.9678 0.9838
No log 63.2 316 0.9755 0.0107 0.9755 0.9877
No log 63.6 318 0.9877 0.0391 0.9877 0.9938
No log 64.0 320 0.9701 0.0107 0.9701 0.9849
No log 64.4 322 0.9382 -0.0211 0.9382 0.9686
No log 64.8 324 0.8911 -0.0500 0.8911 0.9440
No log 65.2 326 0.8526 -0.0393 0.8526 0.9234
No log 65.6 328 0.8417 -0.0376 0.8417 0.9175
No log 66.0 330 0.8409 -0.0376 0.8409 0.9170
No log 66.4 332 0.8628 -0.0471 0.8628 0.9289
No log 66.8 334 0.8846 -0.0854 0.8846 0.9405
No log 67.2 336 0.9197 -0.0194 0.9197 0.9590
No log 67.6 338 0.9238 -0.0194 0.9238 0.9611
No log 68.0 340 0.9043 -0.0526 0.9043 0.9509
No log 68.4 342 0.8924 -0.0157 0.8924 0.9447
No log 68.8 344 0.8711 0.0224 0.8711 0.9333
No log 69.2 346 0.8545 0.0250 0.8545 0.9244
No log 69.6 348 0.8444 0.1065 0.8444 0.9189
No log 70.0 350 0.8512 0.1065 0.8512 0.9226
No log 70.4 352 0.8715 0.0157 0.8715 0.9335
No log 70.8 354 0.8884 0.0805 0.8884 0.9425
No log 71.2 356 0.8837 0.0526 0.8837 0.9401
No log 71.6 358 0.8727 0.0224 0.8727 0.9342
No log 72.0 360 0.8670 0.0651 0.8670 0.9311
No log 72.4 362 0.8671 0.0651 0.8671 0.9312
No log 72.8 364 0.8615 0.0304 0.8615 0.9282
No log 73.2 366 0.8724 0.0304 0.8724 0.9340
No log 73.6 368 0.8780 0.0651 0.8780 0.9370
No log 74.0 370 0.8886 0.0182 0.8886 0.9427
No log 74.4 372 0.9073 0.0839 0.9073 0.9525
No log 74.8 374 0.9085 0.0805 0.9085 0.9532
No log 75.2 376 0.9055 0.0431 0.9055 0.9516
No log 75.6 378 0.8903 0.0515 0.8903 0.9435
No log 76.0 380 0.8662 0.0316 0.8662 0.9307
No log 76.4 382 0.8636 -0.0054 0.8636 0.9293
No log 76.8 384 0.8732 0.0304 0.8732 0.9345
No log 77.2 386 0.8766 0.0304 0.8766 0.9362
No log 77.6 388 0.8821 0.0304 0.8821 0.9392
No log 78.0 390 0.9067 -0.0097 0.9067 0.9522
No log 78.4 392 0.9172 -0.0486 0.9172 0.9577
No log 78.8 394 0.9125 -0.0532 0.9125 0.9553
No log 79.2 396 0.9103 -0.0545 0.9103 0.9541
No log 79.6 398 0.9120 -0.0200 0.9120 0.9550
No log 80.0 400 0.9180 0.0111 0.9180 0.9581
No log 80.4 402 0.9232 0.0111 0.9232 0.9608
No log 80.8 404 0.9105 0.0157 0.9105 0.9542
No log 81.2 406 0.8992 -0.0163 0.8992 0.9483
No log 81.6 408 0.8905 0.0233 0.8905 0.9437
No log 82.0 410 0.8869 -0.0118 0.8869 0.9417
No log 82.4 412 0.8818 -0.0097 0.8818 0.9391
No log 82.8 414 0.8915 -0.0097 0.8915 0.9442
No log 83.2 416 0.8924 -0.0097 0.8924 0.9446
No log 83.6 418 0.8930 0.0304 0.8930 0.9450
No log 84.0 420 0.9057 -0.0097 0.9057 0.9517
No log 84.4 422 0.9135 -0.0500 0.9135 0.9558
No log 84.8 424 0.9064 -0.0097 0.9064 0.9521
No log 85.2 426 0.8964 -0.0097 0.8964 0.9468
No log 85.6 428 0.8906 -0.0097 0.8906 0.9437
No log 86.0 430 0.8866 -0.0097 0.8866 0.9416
No log 86.4 432 0.8797 -0.0097 0.8797 0.9379
No log 86.8 434 0.8798 -0.0486 0.8798 0.9380
No log 87.2 436 0.8822 -0.0513 0.8822 0.9392
No log 87.6 438 0.8921 0.0157 0.8921 0.9445
No log 88.0 440 0.9034 0.0805 0.9034 0.9505
No log 88.4 442 0.9045 0.0805 0.9045 0.9510
No log 88.8 444 0.8977 0.0805 0.8977 0.9475
No log 89.2 446 0.8903 0.0805 0.8903 0.9436
No log 89.6 448 0.8839 0.0805 0.8839 0.9402
No log 90.0 450 0.8703 0.0486 0.8703 0.9329
No log 90.4 452 0.8603 -0.0182 0.8603 0.9275
No log 90.8 454 0.8556 -0.0182 0.8556 0.9250
No log 91.2 456 0.8580 -0.0182 0.8580 0.9263
No log 91.6 458 0.8575 -0.0163 0.8575 0.9260
No log 92.0 460 0.8613 -0.0163 0.8613 0.9281
No log 92.4 462 0.8641 -0.0500 0.8641 0.9296
No log 92.8 464 0.8690 -0.0500 0.8690 0.9322
No log 93.2 466 0.8726 -0.0500 0.8726 0.9341
No log 93.6 468 0.8771 -0.0513 0.8771 0.9365
No log 94.0 470 0.8824 -0.0182 0.8824 0.9394
No log 94.4 472 0.8902 -0.0182 0.8902 0.9435
No log 94.8 474 0.8950 0.0157 0.8950 0.9461
No log 95.2 476 0.8978 0.0157 0.8978 0.9475
No log 95.6 478 0.9006 0.0157 0.9006 0.9490
No log 96.0 480 0.9012 0.0157 0.9012 0.9493
No log 96.4 482 0.9004 0.0157 0.9004 0.9489
No log 96.8 484 0.9001 0.0157 0.9001 0.9487
No log 97.2 486 0.9012 0.0157 0.9012 0.9493
No log 97.6 488 0.9042 0.0157 0.9042 0.9509
No log 98.0 490 0.9077 0.0157 0.9077 0.9527
No log 98.4 492 0.9104 0.0157 0.9104 0.9542
No log 98.8 494 0.9131 0.0486 0.9131 0.9556
No log 99.2 496 0.9154 0.0486 0.9154 0.9568
No log 99.6 498 0.9167 0.0486 0.9167 0.9574
0.1573 100.0 500 0.9171 0.0486 0.9171 0.9577

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k2_task3_organization

Finetuned
(4019)
this model