ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k4_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8550
  • Qwk: -0.0916
  • Mse: 0.8550
  • Rmse: 0.9247

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 4.5110 0.0163 4.5110 2.1239
No log 0.4 4 2.8700 0.0308 2.8700 1.6941
No log 0.6 6 1.5154 0.0444 1.5154 1.2310
No log 0.8 8 2.0564 0.0084 2.0564 1.4340
No log 1.0 10 2.5645 0.0115 2.5645 1.6014
No log 1.2 12 2.1755 -0.0092 2.1755 1.4749
No log 1.4 14 0.7674 -0.0287 0.7674 0.8760
No log 1.6 16 0.6576 0.0 0.6576 0.8109
No log 1.8 18 0.7380 0.0 0.7380 0.8591
No log 2.0 20 0.8484 0.0346 0.8484 0.9211
No log 2.2 22 1.0390 -0.0234 1.0390 1.0193
No log 2.4 24 1.1125 -0.0234 1.1125 1.0547
No log 2.6 26 0.9441 -0.0084 0.9441 0.9717
No log 2.8 28 0.7352 0.0460 0.7352 0.8575
No log 3.0 30 0.7047 0.0506 0.7047 0.8395
No log 3.2 32 0.7370 0.0460 0.7370 0.8585
No log 3.4 34 0.8144 -0.0264 0.8144 0.9024
No log 3.6 36 1.0106 -0.0435 1.0106 1.0053
No log 3.8 38 1.0994 -0.0207 1.0994 1.0485
No log 4.0 40 1.0389 -0.0411 1.0389 1.0193
No log 4.2 42 1.1896 -0.0435 1.1896 1.0907
No log 4.4 44 1.1784 -0.0997 1.1784 1.0855
No log 4.6 46 0.9737 0.0111 0.9737 0.9868
No log 4.8 48 1.0285 0.0046 1.0285 1.0141
No log 5.0 50 0.9702 0.0089 0.9702 0.9850
No log 5.2 52 0.8210 0.0442 0.8210 0.9061
No log 5.4 54 0.7553 0.1449 0.7553 0.8691
No log 5.6 56 0.9214 0.0545 0.9214 0.9599
No log 5.8 58 0.7915 0.1196 0.7915 0.8897
No log 6.0 60 0.6854 0.1379 0.6854 0.8279
No log 6.2 62 0.7735 0.1899 0.7735 0.8795
No log 6.4 64 1.1220 0.0753 1.1220 1.0592
No log 6.6 66 0.7652 0.1573 0.7652 0.8747
No log 6.8 68 0.8511 -0.0616 0.8511 0.9225
No log 7.0 70 0.8701 -0.0970 0.8701 0.9328
No log 7.2 72 0.8239 0.1415 0.8239 0.9077
No log 7.4 74 1.5174 0.0083 1.5174 1.2318
No log 7.6 76 1.4509 -0.0178 1.4509 1.2045
No log 7.8 78 0.9376 -0.0539 0.9376 0.9683
No log 8.0 80 0.7833 -0.0738 0.7833 0.8851
No log 8.2 82 0.7780 -0.0366 0.7780 0.8821
No log 8.4 84 0.8500 0.0062 0.8500 0.9220
No log 8.6 86 1.0577 -0.0169 1.0577 1.0284
No log 8.8 88 1.0547 0.0526 1.0547 1.0270
No log 9.0 90 0.9313 -0.0052 0.9313 0.9650
No log 9.2 92 0.7635 -0.0056 0.7635 0.8738
No log 9.4 94 0.7588 -0.0407 0.7588 0.8711
No log 9.6 96 0.7291 0.0414 0.7291 0.8539
No log 9.8 98 0.8106 0.1342 0.8106 0.9004
No log 10.0 100 0.8033 0.1342 0.8033 0.8963
No log 10.2 102 0.7053 0.1444 0.7053 0.8398
No log 10.4 104 0.7060 0.1021 0.7060 0.8402
No log 10.6 106 0.7709 0.1965 0.7709 0.8780
No log 10.8 108 0.8397 0.0786 0.8397 0.9164
No log 11.0 110 0.7318 0.0909 0.7318 0.8554
No log 11.2 112 0.7636 -0.1329 0.7636 0.8739
No log 11.4 114 0.7833 0.0053 0.7833 0.8850
No log 11.6 116 0.9985 -0.0441 0.9985 0.9993
No log 11.8 118 0.9324 -0.0391 0.9324 0.9656
No log 12.0 120 0.8697 0.0335 0.8697 0.9326
No log 12.2 122 0.9643 0.0164 0.9643 0.9820
No log 12.4 124 0.8485 -0.0015 0.8485 0.9212
No log 12.6 126 0.8690 -0.0008 0.8690 0.9322
No log 12.8 128 1.0882 -0.0586 1.0882 1.0432
No log 13.0 130 0.8713 0.0362 0.8713 0.9334
No log 13.2 132 0.7453 -0.0096 0.7453 0.8633
No log 13.4 134 0.8003 -0.1606 0.8003 0.8946
No log 13.6 136 0.7975 -0.0881 0.7975 0.8930
No log 13.8 138 0.7667 -0.1026 0.7667 0.8756
No log 14.0 140 0.7980 -0.0170 0.7980 0.8933
No log 14.2 142 0.8481 0.0152 0.8481 0.9209
No log 14.4 144 0.7704 0.0395 0.7704 0.8777
No log 14.6 146 0.7935 -0.0939 0.7935 0.8908
No log 14.8 148 0.8419 -0.1753 0.8419 0.9176
No log 15.0 150 0.7805 -0.1001 0.7805 0.8835
No log 15.2 152 0.7530 0.0225 0.7530 0.8677
No log 15.4 154 0.8727 0.0826 0.8727 0.9342
No log 15.6 156 0.8357 0.1342 0.8357 0.9142
No log 15.8 158 0.7564 0.0260 0.7564 0.8697
No log 16.0 160 0.7864 0.0031 0.7864 0.8868
No log 16.2 162 0.8125 -0.0366 0.8125 0.9014
No log 16.4 164 0.8071 -0.0595 0.8071 0.8984
No log 16.6 166 0.8371 0.1047 0.8371 0.9149
No log 16.8 168 0.9713 -0.0182 0.9713 0.9855
No log 17.0 170 0.9299 -0.0163 0.9299 0.9643
No log 17.2 172 0.7823 0.0282 0.7823 0.8845
No log 17.4 174 0.7614 -0.0062 0.7614 0.8726
No log 17.6 176 0.7589 0.0414 0.7589 0.8711
No log 17.8 178 0.7686 0.0814 0.7686 0.8767
No log 18.0 180 0.7926 0.1199 0.7926 0.8903
No log 18.2 182 0.7987 0.1254 0.7987 0.8937
No log 18.4 184 0.7882 0.0375 0.7882 0.8878
No log 18.6 186 0.7710 -0.0493 0.7710 0.8781
No log 18.8 188 0.7395 0.0061 0.7395 0.8600
No log 19.0 190 0.7297 0.0282 0.7297 0.8543
No log 19.2 192 0.7397 0.0723 0.7397 0.8601
No log 19.4 194 0.7500 0.0214 0.7500 0.8661
No log 19.6 196 0.6987 -0.0032 0.6987 0.8359
No log 19.8 198 0.7058 0.0479 0.7058 0.8401
No log 20.0 200 0.7776 0.0562 0.7776 0.8818
No log 20.2 202 0.8511 -0.0845 0.8511 0.9225
No log 20.4 204 0.7924 0.1440 0.7924 0.8902
No log 20.6 206 0.7143 -0.0033 0.7143 0.8452
No log 20.8 208 0.7193 -0.0451 0.7193 0.8481
No log 21.0 210 0.6956 0.1021 0.6956 0.8341
No log 21.2 212 0.8050 0.1291 0.8050 0.8972
No log 21.4 214 0.9865 -0.0923 0.9865 0.9932
No log 21.6 216 0.9159 -0.0518 0.9159 0.9570
No log 21.8 218 0.7962 0.0247 0.7962 0.8923
No log 22.0 220 0.7623 -0.0541 0.7623 0.8731
No log 22.2 222 0.7651 -0.0541 0.7651 0.8747
No log 22.4 224 0.8002 0.0869 0.8002 0.8946
No log 22.6 226 0.8053 -0.0195 0.8053 0.8974
No log 22.8 228 0.7856 0.0869 0.7856 0.8864
No log 23.0 230 0.7547 0.0914 0.7547 0.8688
No log 23.2 232 0.7430 0.0436 0.7430 0.8620
No log 23.4 234 0.7328 0.0 0.7328 0.8560
No log 23.6 236 0.7413 0.0436 0.7413 0.8610
No log 23.8 238 0.8711 -0.0878 0.8711 0.9333
No log 24.0 240 1.0079 0.0353 1.0079 1.0040
No log 24.2 242 0.9327 -0.0583 0.9327 0.9658
No log 24.4 244 0.7677 0.0680 0.7677 0.8762
No log 24.6 246 0.7308 0.0914 0.7308 0.8549
No log 24.8 248 0.7281 0.0094 0.7281 0.8533
No log 25.0 250 0.7254 0.1024 0.7254 0.8517
No log 25.2 252 0.7802 -0.0812 0.7802 0.8833
No log 25.4 254 0.7687 -0.0351 0.7687 0.8767
No log 25.6 256 0.7161 0.0460 0.7161 0.8462
No log 25.8 258 0.7513 -0.0428 0.7513 0.8668
No log 26.0 260 0.8103 -0.1201 0.8103 0.9001
No log 26.2 262 0.8509 -0.1033 0.8509 0.9224
No log 26.4 264 0.7596 0.0628 0.7596 0.8716
No log 26.6 266 0.7145 -0.0065 0.7145 0.8453
No log 26.8 268 0.7677 0.0723 0.7677 0.8762
No log 27.0 270 0.7786 0.0680 0.7786 0.8824
No log 27.2 272 0.7937 0.0723 0.7937 0.8909
No log 27.4 274 0.7297 -0.0560 0.7297 0.8542
No log 27.6 276 0.7240 -0.0062 0.7240 0.8509
No log 27.8 278 0.7155 -0.0062 0.7155 0.8459
No log 28.0 280 0.7106 0.0436 0.7106 0.8429
No log 28.2 282 0.7323 0.0863 0.7323 0.8557
No log 28.4 284 0.7343 0.0318 0.7343 0.8569
No log 28.6 286 0.7227 0.0914 0.7227 0.8501
No log 28.8 288 0.7377 0.0318 0.7377 0.8589
No log 29.0 290 0.7515 0.0680 0.7515 0.8669
No log 29.2 292 0.7246 0.0355 0.7246 0.8513
No log 29.4 294 0.7149 0.0914 0.7149 0.8455
No log 29.6 296 0.7106 0.0436 0.7106 0.8430
No log 29.8 298 0.7056 0.0964 0.7056 0.8400
No log 30.0 300 0.7065 0.0964 0.7065 0.8405
No log 30.2 302 0.7262 0.1202 0.7262 0.8522
No log 30.4 304 0.7204 0.0909 0.7204 0.8488
No log 30.6 306 0.7206 0.0436 0.7206 0.8489
No log 30.8 308 0.7245 0.0970 0.7245 0.8512
No log 31.0 310 0.7406 0.0741 0.7406 0.8606
No log 31.2 312 0.7689 0.0660 0.7689 0.8769
No log 31.4 314 0.7729 0.0660 0.7729 0.8791
No log 31.6 316 0.7552 0.0660 0.7552 0.8690
No log 31.8 318 0.7618 0.0639 0.7618 0.8728
No log 32.0 320 0.7418 0.0639 0.7418 0.8613
No log 32.2 322 0.7047 0.1371 0.7047 0.8395
No log 32.4 324 0.7034 0.1023 0.7034 0.8387
No log 32.6 326 0.7058 0.0556 0.7058 0.8401
No log 32.8 328 0.7098 0.0375 0.7098 0.8425
No log 33.0 330 0.7322 -0.0170 0.7322 0.8557
No log 33.2 332 0.7709 0.0588 0.7709 0.8780
No log 33.4 334 0.7913 0.0525 0.7913 0.8895
No log 33.6 336 0.8692 0.0182 0.8692 0.9323
No log 33.8 338 0.8821 0.0182 0.8821 0.9392
No log 34.0 340 0.8131 -0.0408 0.8131 0.9017
No log 34.2 342 0.7517 0.0375 0.7517 0.8670
No log 34.4 344 0.7656 0.0033 0.7656 0.8750
No log 34.6 346 0.8160 -0.0591 0.8160 0.9033
No log 34.8 348 0.8298 -0.1474 0.8298 0.9109
No log 35.0 350 0.7881 -0.1142 0.7881 0.8877
No log 35.2 352 0.7219 0.0064 0.7219 0.8497
No log 35.4 354 0.7248 0.0479 0.7248 0.8514
No log 35.6 356 0.8669 0.0111 0.8669 0.9311
No log 35.8 358 0.9846 0.0196 0.9846 0.9923
No log 36.0 360 0.9891 0.0305 0.9891 0.9945
No log 36.2 362 0.8313 0.0642 0.8313 0.9118
No log 36.4 364 0.7498 0.0680 0.7498 0.8659
No log 36.6 366 0.7439 0.0 0.7439 0.8625
No log 36.8 368 0.7526 0.0 0.7526 0.8675
No log 37.0 370 0.7469 0.0375 0.7469 0.8642
No log 37.2 372 0.7714 -0.0295 0.7714 0.8783
No log 37.4 374 0.8074 -0.0373 0.8074 0.8986
No log 37.6 376 0.8310 -0.0809 0.8310 0.9116
No log 37.8 378 0.8041 -0.0755 0.8041 0.8967
No log 38.0 380 0.7781 -0.0331 0.7781 0.8821
No log 38.2 382 0.7725 0.0588 0.7725 0.8789
No log 38.4 384 0.7360 0.0759 0.7360 0.8579
No log 38.6 386 0.7234 0.0807 0.7234 0.8505
No log 38.8 388 0.7257 0.0318 0.7257 0.8519
No log 39.0 390 0.7307 0.0318 0.7307 0.8548
No log 39.2 392 0.7559 0.0680 0.7559 0.8694
No log 39.4 394 0.7931 0.0512 0.7931 0.8906
No log 39.6 396 0.7968 0.0512 0.7968 0.8926
No log 39.8 398 0.7824 0.0549 0.7824 0.8845
No log 40.0 400 0.7857 0.0549 0.7857 0.8864
No log 40.2 402 0.7688 0.0247 0.7688 0.8768
No log 40.4 404 0.7515 0.0479 0.7515 0.8669
No log 40.6 406 0.7499 0.0479 0.7499 0.8660
No log 40.8 408 0.7557 0.0967 0.7557 0.8693
No log 41.0 410 0.7456 0.0436 0.7456 0.8635
No log 41.2 412 0.7655 0.0680 0.7655 0.8749
No log 41.4 414 0.8276 0.0017 0.8276 0.9097
No log 41.6 416 0.8706 -0.0504 0.8706 0.9331
No log 41.8 418 0.8360 -0.0033 0.8360 0.9143
No log 42.0 420 0.8056 0.0476 0.8056 0.8975
No log 42.2 422 0.7567 0.0768 0.7567 0.8699
No log 42.4 424 0.7535 0.0436 0.7535 0.8681
No log 42.6 426 0.7571 -0.0096 0.7571 0.8701
No log 42.8 428 0.7452 -0.0541 0.7452 0.8632
No log 43.0 430 0.7535 -0.0091 0.7535 0.8680
No log 43.2 432 0.7870 0.0318 0.7870 0.8871
No log 43.4 434 0.8698 -0.1224 0.8698 0.9326
No log 43.6 436 0.9308 -0.0955 0.9308 0.9648
No log 43.8 438 0.9314 -0.0918 0.9314 0.9651
No log 44.0 440 0.8628 -0.1221 0.8628 0.9289
No log 44.2 442 0.8451 -0.1547 0.8451 0.9193
No log 44.4 444 0.8194 -0.0240 0.8194 0.9052
No log 44.6 446 0.8070 0.0236 0.8070 0.8983
No log 44.8 448 0.8147 0.0175 0.8147 0.9026
No log 45.0 450 0.8350 -0.0809 0.8350 0.9138
No log 45.2 452 0.8487 -0.1212 0.8487 0.9213
No log 45.4 454 0.8463 -0.1212 0.8463 0.9199
No log 45.6 456 0.8859 -0.1271 0.8859 0.9412
No log 45.8 458 0.8506 -0.1267 0.8506 0.9223
No log 46.0 460 0.8131 -0.0425 0.8131 0.9017
No log 46.2 462 0.7674 0.0768 0.7674 0.8760
No log 46.4 464 0.7528 0.0355 0.7528 0.8677
No log 46.6 466 0.7540 0.0338 0.7540 0.8684
No log 46.8 468 0.7554 0.0776 0.7554 0.8691
No log 47.0 470 0.7521 0.1096 0.7521 0.8673
No log 47.2 472 0.7419 0.1146 0.7419 0.8613
No log 47.4 474 0.7322 0.0454 0.7322 0.8557
No log 47.6 476 0.7342 0.0436 0.7342 0.8569
No log 47.8 478 0.7446 0.1254 0.7446 0.8629
No log 48.0 480 0.7758 0.0588 0.7758 0.8808
No log 48.2 482 0.8324 -0.0788 0.8324 0.9124
No log 48.4 484 0.8837 -0.1589 0.8837 0.9400
No log 48.6 486 0.9233 -0.1230 0.9233 0.9609
No log 48.8 488 0.8670 -0.1605 0.8670 0.9311
No log 49.0 490 0.7982 0.1202 0.7982 0.8934
No log 49.2 492 0.7860 0.1691 0.7860 0.8866
No log 49.4 494 0.8120 -0.0408 0.8120 0.9011
No log 49.6 496 0.8602 -0.1271 0.8602 0.9275
No log 49.8 498 0.9044 -0.1274 0.9044 0.9510
0.2396 50.0 500 0.9461 -0.0606 0.9461 0.9727
0.2396 50.2 502 0.9218 -0.1276 0.9218 0.9601
0.2396 50.4 504 0.8569 -0.1957 0.8569 0.9257
0.2396 50.6 506 0.8241 -0.0295 0.8241 0.9078
0.2396 50.8 508 0.8350 -0.1200 0.8350 0.9138
0.2396 51.0 510 0.8146 0.0282 0.8146 0.9026
0.2396 51.2 512 0.7759 0.0395 0.7759 0.8809
0.2396 51.4 514 0.7634 -0.0091 0.7634 0.8737
0.2396 51.6 516 0.7683 -0.0091 0.7683 0.8765
0.2396 51.8 518 0.7856 -0.0612 0.7856 0.8864
0.2396 52.0 520 0.8391 -0.0390 0.8391 0.9160
0.2396 52.2 522 0.8640 -0.1265 0.8640 0.9295
0.2396 52.4 524 0.8695 -0.0916 0.8695 0.9325
0.2396 52.6 526 0.8902 -0.1277 0.8902 0.9435
0.2396 52.8 528 0.8739 -0.0923 0.8739 0.9348
0.2396 53.0 530 0.8550 -0.0916 0.8550 0.9247

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingWellWrittenEssays_FineTuningAraBERT_run2_AugV5_k4_task3_organization

Finetuned
(4019)
this model