ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k11_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.3032
  • Qwk: 0.1125
  • Mse: 1.3032
  • Rmse: 1.1416

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0408 2 3.3201 -0.0350 3.3201 1.8221
No log 0.0816 4 1.6711 -0.0130 1.6711 1.2927
No log 0.1224 6 1.2888 0.0335 1.2888 1.1353
No log 0.1633 8 0.6280 -0.0411 0.6280 0.7924
No log 0.2041 10 0.6590 -0.0159 0.6590 0.8118
No log 0.2449 12 0.7494 -0.0732 0.7494 0.8657
No log 0.2857 14 0.7579 -0.0746 0.7579 0.8706
No log 0.3265 16 0.7254 -0.0909 0.7254 0.8517
No log 0.3673 18 0.6949 0.0476 0.6949 0.8336
No log 0.4082 20 0.7215 -0.0159 0.7215 0.8494
No log 0.4490 22 0.6605 -0.0081 0.6605 0.8127
No log 0.4898 24 0.6185 0.0 0.6185 0.7864
No log 0.5306 26 0.6179 0.0 0.6179 0.7860
No log 0.5714 28 0.6212 0.0 0.6212 0.7882
No log 0.6122 30 0.6297 0.0 0.6297 0.7935
No log 0.6531 32 0.7004 0.0 0.7004 0.8369
No log 0.6939 34 0.6817 0.0 0.6817 0.8257
No log 0.7347 36 0.6374 0.0 0.6374 0.7984
No log 0.7755 38 0.6924 -0.0963 0.6924 0.8321
No log 0.8163 40 0.7198 -0.1111 0.7198 0.8484
No log 0.8571 42 0.7251 -0.1631 0.7251 0.8515
No log 0.8980 44 0.6838 -0.0963 0.6838 0.8269
No log 0.9388 46 0.6605 -0.0233 0.6605 0.8127
No log 0.9796 48 0.6493 -0.0159 0.6493 0.8058
No log 1.0204 50 0.7023 0.0720 0.7023 0.8380
No log 1.0612 52 0.7156 0.0769 0.7156 0.8460
No log 1.1020 54 0.6620 0.0067 0.6620 0.8136
No log 1.1429 56 0.7612 0.0647 0.7612 0.8724
No log 1.1837 58 0.8426 0.0370 0.8426 0.9179
No log 1.2245 60 0.6950 0.0380 0.6950 0.8337
No log 1.2653 62 0.7629 0.1195 0.7629 0.8735
No log 1.3061 64 0.7838 0.1285 0.7838 0.8853
No log 1.3469 66 0.7048 0.1325 0.7048 0.8395
No log 1.3878 68 0.7108 0.1138 0.7108 0.8431
No log 1.4286 70 0.7607 0.0703 0.7607 0.8722
No log 1.4694 72 0.6818 0.0526 0.6818 0.8257
No log 1.5102 74 0.7421 0.0545 0.7421 0.8615
No log 1.5510 76 0.7816 0.0638 0.7816 0.8841
No log 1.5918 78 1.0772 0.0866 1.0772 1.0379
No log 1.6327 80 1.0615 0.0551 1.0615 1.0303
No log 1.6735 82 0.8017 0.0374 0.8017 0.8954
No log 1.7143 84 0.8436 0.0457 0.8436 0.9185
No log 1.7551 86 0.8352 0.0157 0.8352 0.9139
No log 1.7959 88 0.8219 0.0359 0.8219 0.9066
No log 1.8367 90 1.1136 0.0270 1.1136 1.0553
No log 1.8776 92 1.1101 0.0551 1.1101 1.0536
No log 1.9184 94 1.0601 0.0820 1.0601 1.0296
No log 1.9592 96 0.9032 0.0049 0.9032 0.9504
No log 2.0 98 0.9124 0.0316 0.9124 0.9552
No log 2.0408 100 0.8845 0.125 0.8845 0.9405
No log 2.0816 102 0.9243 0.0 0.9243 0.9614
No log 2.1224 104 1.2570 0.0111 1.2570 1.1212
No log 2.1633 106 1.2108 0.0335 1.2108 1.1004
No log 2.2041 108 0.9503 0.0230 0.9503 0.9748
No log 2.2449 110 0.8059 0.1269 0.8059 0.8977
No log 2.2857 112 0.7734 0.1304 0.7734 0.8794
No log 2.3265 114 0.7662 0.1158 0.7662 0.8753
No log 2.3673 116 0.7733 0.0497 0.7733 0.8794
No log 2.4082 118 0.8000 0.0439 0.8000 0.8944
No log 2.4490 120 0.8683 0.0512 0.8683 0.9318
No log 2.4898 122 0.9555 0.0968 0.9555 0.9775
No log 2.5306 124 1.0150 0.0207 1.0150 1.0075
No log 2.5714 126 1.1068 0.1818 1.1068 1.0520
No log 2.6122 128 1.4303 0.1329 1.4303 1.1959
No log 2.6531 130 1.6649 0.1160 1.6649 1.2903
No log 2.6939 132 1.1444 0.1533 1.1444 1.0698
No log 2.7347 134 0.9507 0.1807 0.9507 0.9750
No log 2.7755 136 0.9122 0.2510 0.9122 0.9551
No log 2.8163 138 1.3730 0.1429 1.3730 1.1717
No log 2.8571 140 2.3968 0.0619 2.3968 1.5482
No log 2.8980 142 2.3368 0.0427 2.3368 1.5287
No log 2.9388 144 1.6572 0.1304 1.6572 1.2873
No log 2.9796 146 1.3528 0.0809 1.3528 1.1631
No log 3.0204 148 1.5537 0.1097 1.5537 1.2465
No log 3.0612 150 2.0841 0.0386 2.0841 1.4437
No log 3.1020 152 2.0331 0.0148 2.0331 1.4259
No log 3.1429 154 1.7148 0.0284 1.7148 1.3095
No log 3.1837 156 1.6881 0.0573 1.6881 1.2993
No log 3.2245 158 1.8515 -0.0029 1.8515 1.3607
No log 3.2653 160 1.8323 0.0247 1.8323 1.3536
No log 3.3061 162 1.8468 0.0227 1.8468 1.3590
No log 3.3469 164 1.5144 0.0789 1.5144 1.2306
No log 3.3878 166 1.3421 0.1143 1.3421 1.1585
No log 3.4286 168 1.5902 0.0062 1.5902 1.2610
No log 3.4694 170 1.6753 -0.0265 1.6753 1.2944
No log 3.5102 172 2.0379 0.0392 2.0379 1.4276
No log 3.5510 174 1.9038 0.0028 1.9038 1.3798
No log 3.5918 176 1.8042 -0.0233 1.8042 1.3432
No log 3.6327 178 1.9530 0.0313 1.9530 1.3975
No log 3.6735 180 2.0663 0.0028 2.0663 1.4375
No log 3.7143 182 1.6946 0.0289 1.6946 1.3018
No log 3.7551 184 1.6550 0.0264 1.6550 1.2865
No log 3.7959 186 2.0208 -0.0221 2.0208 1.4216
No log 3.8367 188 2.0989 -0.0479 2.0989 1.4488
No log 3.8776 190 2.0212 -0.0029 2.0212 1.4217
No log 3.9184 192 1.7547 -0.0149 1.7547 1.3246
No log 3.9592 194 1.3094 0.1095 1.3094 1.1443
No log 4.0 196 1.2511 0.1292 1.2511 1.1185
No log 4.0408 198 1.4935 0.0476 1.4935 1.2221
No log 4.0816 200 1.9928 0.0437 1.9928 1.4117
No log 4.1224 202 1.9347 0.0142 1.9347 1.3909
No log 4.1633 204 1.4739 0.0405 1.4739 1.2140
No log 4.2041 206 1.0624 0.0871 1.0624 1.0307
No log 4.2449 208 1.0740 0.0871 1.0740 1.0364
No log 4.2857 210 1.4210 0.0886 1.4210 1.1921
No log 4.3265 212 1.9922 0.0457 1.9922 1.4115
No log 4.3673 214 2.0209 0.0515 2.0209 1.4216
No log 4.4082 216 1.5660 0.0 1.5660 1.2514
No log 4.4490 218 1.3760 0.0891 1.3760 1.1730
No log 4.4898 220 1.2072 0.1014 1.2072 1.0987
No log 4.5306 222 1.4147 0.0868 1.4147 1.1894
No log 4.5714 224 1.4427 0.1083 1.4427 1.2011
No log 4.6122 226 1.3381 0.1316 1.3381 1.1568
No log 4.6531 228 1.5191 0.0843 1.5191 1.2325
No log 4.6939 230 1.6444 0.0659 1.6444 1.2824
No log 4.7347 232 2.0807 0.0420 2.0807 1.4425
No log 4.7755 234 2.1061 0.0449 2.1061 1.4512
No log 4.8163 236 1.9599 0.0408 1.9599 1.4000
No log 4.8571 238 1.5628 0.0415 1.5628 1.2501
No log 4.8980 240 1.4542 0.0915 1.4542 1.2059
No log 4.9388 242 1.6683 -0.0182 1.6683 1.2916
No log 4.9796 244 1.9313 0.0866 1.9313 1.3897
No log 5.0204 246 1.8003 0.0423 1.8003 1.3418
No log 5.0612 248 1.6350 -0.0210 1.6350 1.2787
No log 5.1020 250 1.4113 0.1141 1.4113 1.1880
No log 5.1429 252 1.2631 0.0556 1.2631 1.1239
No log 5.1837 254 1.2682 0.0580 1.2682 1.1261
No log 5.2245 256 1.5676 0.0984 1.5676 1.2520
No log 5.2653 258 1.9882 -0.0213 1.9882 1.4100
No log 5.3061 260 1.9162 -0.0028 1.9162 1.3843
No log 5.3469 262 1.5754 0.0523 1.5754 1.2552
No log 5.3878 264 1.2260 0.1081 1.2260 1.1073
No log 5.4286 266 1.2173 0.1608 1.2173 1.1033
No log 5.4694 268 1.1956 0.1317 1.1956 1.0934
No log 5.5102 270 1.3481 0.1373 1.3481 1.1611
No log 5.5510 272 1.4990 0.1226 1.4990 1.2243
No log 5.5918 274 1.4453 0.0933 1.4453 1.2022
No log 5.6327 276 1.4124 0.1111 1.4124 1.1885
No log 5.6735 278 1.2907 0.1081 1.2907 1.1361
No log 5.7143 280 1.3526 0.1081 1.3526 1.1630
No log 5.7551 282 1.6427 0.0991 1.6427 1.2817
No log 5.7959 284 1.8781 -0.0305 1.8781 1.3704
No log 5.8367 286 1.7632 -0.0115 1.7632 1.3278
No log 5.8776 288 1.3699 0.1068 1.3699 1.1704
No log 5.9184 290 1.1192 0.1292 1.1192 1.0579
No log 5.9592 292 1.0849 0.1292 1.0849 1.0416
No log 6.0 294 1.2426 0.0565 1.2426 1.1147
No log 6.0408 296 1.6209 0.0877 1.6209 1.2731
No log 6.0816 298 1.7746 0.0508 1.7746 1.3321
No log 6.1224 300 1.6602 -0.0029 1.6602 1.2885
No log 6.1633 302 1.4069 0.0886 1.4069 1.1861
No log 6.2041 304 1.4091 0.1083 1.4091 1.1871
No log 6.2449 306 1.6671 0.0734 1.6671 1.2911
No log 6.2857 308 2.0108 0.0549 2.0108 1.4180
No log 6.3265 310 1.9321 0.0496 1.9321 1.3900
No log 6.3673 312 1.7453 0.0 1.7453 1.3211
No log 6.4082 314 1.6025 0.0734 1.6025 1.2659
No log 6.4490 316 1.7310 0.0276 1.7310 1.3157
No log 6.4898 318 1.7196 0.0029 1.7196 1.3113
No log 6.5306 320 1.4858 0.1041 1.4858 1.2189
No log 6.5714 322 1.3548 0.1139 1.3548 1.1639
No log 6.6122 324 1.3201 0.1125 1.3201 1.1489
No log 6.6531 326 1.4650 0.0968 1.4650 1.2104
No log 6.6939 328 1.4890 0.0927 1.4890 1.2202
No log 6.7347 330 1.5612 0.0250 1.5612 1.2495
No log 6.7755 332 1.5279 0.0927 1.5279 1.2361
No log 6.8163 334 1.4032 0.1139 1.4032 1.1846
No log 6.8571 336 1.4453 0.1139 1.4453 1.2022
No log 6.8980 338 1.4677 0.0927 1.4677 1.2115
No log 6.9388 340 1.5881 -0.0182 1.5881 1.2602
No log 6.9796 342 1.5579 -0.0182 1.5579 1.2481
No log 7.0204 344 1.4619 0.0712 1.4619 1.2091
No log 7.0612 346 1.3970 0.1169 1.3970 1.1819
No log 7.1020 348 1.4363 0.0712 1.4363 1.1984
No log 7.1429 350 1.3777 0.0927 1.3777 1.1737
No log 7.1837 352 1.2986 0.1125 1.2986 1.1396
No log 7.2245 354 1.3867 0.0927 1.3867 1.1776
No log 7.2653 356 1.5210 -0.0182 1.5210 1.2333
No log 7.3061 358 1.4832 0.0244 1.4832 1.2179
No log 7.3469 360 1.4765 0.0464 1.4765 1.2151
No log 7.3878 362 1.4624 0.0464 1.4624 1.2093
No log 7.4286 364 1.3311 0.1139 1.3311 1.1537
No log 7.4694 366 1.2459 0.0831 1.2459 1.1162
No log 7.5102 368 1.2007 0.1317 1.2007 1.0958
No log 7.5510 370 1.2633 0.0850 1.2633 1.1240
No log 7.5918 372 1.4701 0.0270 1.4701 1.2125
No log 7.6327 374 1.5951 0.0296 1.5951 1.2630
No log 7.6735 376 1.5678 0.0296 1.5678 1.2521
No log 7.7143 378 1.4266 0.1111 1.4266 1.1944
No log 7.7551 380 1.2607 0.1081 1.2607 1.1228
No log 7.7959 382 1.2216 0.1329 1.2216 1.1052
No log 7.8367 384 1.2529 0.1329 1.2529 1.1193
No log 7.8776 386 1.3633 0.1111 1.3633 1.1676
No log 7.9184 388 1.5367 0.0453 1.5367 1.2397
No log 7.9592 390 1.5973 0.0520 1.5973 1.2638
No log 8.0 392 1.5930 0.0733 1.5930 1.2621
No log 8.0408 394 1.4835 0.0488 1.4835 1.2180
No log 8.0816 396 1.4354 0.0960 1.4354 1.1981
No log 8.1224 398 1.4255 0.0960 1.4255 1.1939
No log 8.1633 400 1.3677 0.0927 1.3677 1.1695
No log 8.2041 402 1.3612 0.0927 1.3612 1.1667
No log 8.2449 404 1.4209 0.0920 1.4209 1.1920
No log 8.2857 406 1.4605 0.0920 1.4605 1.2085
No log 8.3265 408 1.4548 0.0920 1.4548 1.2062
No log 8.3673 410 1.4614 0.1125 1.4614 1.2089
No log 8.4082 412 1.5265 0.0920 1.5265 1.2355
No log 8.4490 414 1.6114 0.0698 1.6114 1.2694
No log 8.4898 416 1.6261 0.0264 1.6261 1.2752
No log 8.5306 418 1.6115 0.0264 1.6115 1.2694
No log 8.5714 420 1.5824 0.0264 1.5824 1.2579
No log 8.6122 422 1.4735 0.0927 1.4735 1.2139
No log 8.6531 424 1.3534 0.1125 1.3534 1.1633
No log 8.6939 426 1.2796 0.0831 1.2796 1.1312
No log 8.7347 428 1.2783 0.0831 1.2783 1.1306
No log 8.7755 430 1.3142 0.1125 1.3142 1.1464
No log 8.8163 432 1.4021 0.0927 1.4021 1.1841
No log 8.8571 434 1.5053 0.0244 1.5053 1.2269
No log 8.8980 436 1.5556 0.0328 1.5556 1.2473
No log 8.9388 438 1.5160 0.0277 1.5160 1.2313
No log 8.9796 440 1.4311 0.0440 1.4311 1.1963
No log 9.0204 442 1.3219 0.0909 1.3219 1.1497
No log 9.0612 444 1.2447 0.1096 1.2447 1.1157
No log 9.1020 446 1.1993 0.1081 1.1993 1.0951
No log 9.1429 448 1.1999 0.1081 1.1999 1.0954
No log 9.1837 450 1.2125 0.1081 1.2125 1.1011
No log 9.2245 452 1.2643 0.1096 1.2643 1.1244
No log 9.2653 454 1.3394 0.0909 1.3394 1.1573
No log 9.3061 456 1.4354 0.0440 1.4354 1.1981
No log 9.3469 458 1.4882 0.0453 1.4882 1.2199
No log 9.3878 460 1.4943 0.0453 1.4943 1.2224
No log 9.4286 462 1.4708 0.0453 1.4708 1.2127
No log 9.4694 464 1.4314 0.0927 1.4314 1.1964
No log 9.5102 466 1.4040 0.0927 1.4040 1.1849
No log 9.5510 468 1.3811 0.0909 1.3811 1.1752
No log 9.5918 470 1.3543 0.1125 1.3543 1.1637
No log 9.6327 472 1.3483 0.1125 1.3483 1.1612
No log 9.6735 474 1.3393 0.1125 1.3393 1.1573
No log 9.7143 476 1.3253 0.1125 1.3253 1.1512
No log 9.7551 478 1.3246 0.1125 1.3246 1.1509
No log 9.7959 480 1.3259 0.1125 1.3259 1.1515
No log 9.8367 482 1.3175 0.1125 1.3175 1.1478
No log 9.8776 484 1.3132 0.1125 1.3132 1.1460
No log 9.9184 486 1.3092 0.1125 1.3092 1.1442
No log 9.9592 488 1.3047 0.1125 1.3047 1.1423
No log 10.0 490 1.3032 0.1125 1.3032 1.1416

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
1
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run2_AugV5_k11_task3_organization

Finetuned
(4023)
this model