ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k11_task3_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 1.2996
  • Qwk: 0.1125
  • Mse: 1.2996
  • Rmse: 1.1400

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0408 2 3.3201 -0.0350 3.3201 1.8221
No log 0.0816 4 1.6711 -0.0130 1.6711 1.2927
No log 0.1224 6 1.2888 0.0335 1.2888 1.1353
No log 0.1633 8 0.6280 -0.0411 0.6280 0.7924
No log 0.2041 10 0.6590 -0.0159 0.6590 0.8118
No log 0.2449 12 0.7494 -0.0732 0.7494 0.8657
No log 0.2857 14 0.7579 -0.0746 0.7579 0.8706
No log 0.3265 16 0.7254 -0.0909 0.7254 0.8517
No log 0.3673 18 0.6949 0.0476 0.6949 0.8336
No log 0.4082 20 0.7215 -0.0159 0.7215 0.8494
No log 0.4490 22 0.6605 -0.0081 0.6605 0.8127
No log 0.4898 24 0.6185 0.0 0.6185 0.7864
No log 0.5306 26 0.6179 0.0 0.6179 0.7860
No log 0.5714 28 0.6212 0.0 0.6212 0.7882
No log 0.6122 30 0.6297 0.0 0.6297 0.7935
No log 0.6531 32 0.7004 0.0 0.7004 0.8369
No log 0.6939 34 0.6817 0.0 0.6817 0.8257
No log 0.7347 36 0.6374 0.0 0.6374 0.7984
No log 0.7755 38 0.6924 -0.0963 0.6924 0.8321
No log 0.8163 40 0.7198 -0.1111 0.7198 0.8484
No log 0.8571 42 0.7251 -0.1631 0.7251 0.8515
No log 0.8980 44 0.6838 -0.0963 0.6838 0.8269
No log 0.9388 46 0.6605 -0.0233 0.6605 0.8127
No log 0.9796 48 0.6493 -0.0159 0.6493 0.8058
No log 1.0204 50 0.7023 0.0720 0.7023 0.8380
No log 1.0612 52 0.7156 0.0769 0.7156 0.8460
No log 1.1020 54 0.6620 0.0067 0.6620 0.8136
No log 1.1429 56 0.7612 0.0647 0.7612 0.8724
No log 1.1837 58 0.8426 0.0370 0.8426 0.9179
No log 1.2245 60 0.6950 0.0380 0.6950 0.8337
No log 1.2653 62 0.7629 0.1195 0.7629 0.8735
No log 1.3061 64 0.7838 0.1285 0.7838 0.8853
No log 1.3469 66 0.7048 0.1325 0.7048 0.8395
No log 1.3878 68 0.7108 0.1138 0.7108 0.8431
No log 1.4286 70 0.7606 0.0703 0.7606 0.8722
No log 1.4694 72 0.6817 0.0526 0.6817 0.8257
No log 1.5102 74 0.7421 0.0545 0.7421 0.8615
No log 1.5510 76 0.7816 0.0638 0.7816 0.8841
No log 1.5918 78 1.0772 0.0866 1.0772 1.0379
No log 1.6327 80 1.0615 0.0551 1.0615 1.0303
No log 1.6735 82 0.8017 0.0374 0.8017 0.8954
No log 1.7143 84 0.8436 0.0457 0.8436 0.9185
No log 1.7551 86 0.8352 0.0157 0.8352 0.9139
No log 1.7959 88 0.8219 0.0359 0.8219 0.9066
No log 1.8367 90 1.1136 0.0270 1.1136 1.0553
No log 1.8776 92 1.1101 0.0551 1.1101 1.0536
No log 1.9184 94 1.0602 0.0820 1.0602 1.0296
No log 1.9592 96 0.9032 0.0049 0.9032 0.9504
No log 2.0 98 0.9125 0.0316 0.9125 0.9552
No log 2.0408 100 0.8845 0.125 0.8845 0.9405
No log 2.0816 102 0.9243 0.0 0.9243 0.9614
No log 2.1224 104 1.2570 0.0111 1.2570 1.1212
No log 2.1633 106 1.2107 0.0335 1.2107 1.1003
No log 2.2041 108 0.9502 0.0230 0.9502 0.9748
No log 2.2449 110 0.8059 0.1269 0.8059 0.8977
No log 2.2857 112 0.7734 0.1304 0.7734 0.8794
No log 2.3265 114 0.7663 0.1158 0.7663 0.8754
No log 2.3673 116 0.7733 0.0497 0.7733 0.8794
No log 2.4082 118 0.7999 0.0439 0.7999 0.8944
No log 2.4490 120 0.8682 0.0512 0.8682 0.9318
No log 2.4898 122 0.9555 0.0968 0.9555 0.9775
No log 2.5306 124 1.0150 0.0207 1.0150 1.0075
No log 2.5714 126 1.1067 0.1818 1.1067 1.0520
No log 2.6122 128 1.4300 0.1329 1.4300 1.1958
No log 2.6531 130 1.6648 0.1160 1.6648 1.2903
No log 2.6939 132 1.1445 0.1533 1.1445 1.0698
No log 2.7347 134 0.9508 0.1807 0.9508 0.9751
No log 2.7755 136 0.9123 0.2510 0.9123 0.9552
No log 2.8163 138 1.3735 0.1429 1.3735 1.1720
No log 2.8571 140 2.3971 0.0619 2.3971 1.5483
No log 2.8980 142 2.3368 0.0427 2.3368 1.5286
No log 2.9388 144 1.6570 0.1304 1.6570 1.2873
No log 2.9796 146 1.3530 0.0809 1.3530 1.1632
No log 3.0204 148 1.5541 0.1097 1.5541 1.2466
No log 3.0612 150 2.0849 0.0386 2.0849 1.4439
No log 3.1020 152 2.0337 0.0148 2.0337 1.4261
No log 3.1429 154 1.7145 0.0284 1.7145 1.3094
No log 3.1837 156 1.6875 0.0573 1.6875 1.2990
No log 3.2245 158 1.8512 -0.0029 1.8512 1.3606
No log 3.2653 160 1.8329 0.0247 1.8329 1.3538
No log 3.3061 162 1.8482 0.0227 1.8482 1.3595
No log 3.3469 164 1.5150 0.0789 1.5150 1.2309
No log 3.3878 166 1.3418 0.1143 1.3418 1.1583
No log 3.4286 168 1.5894 0.0062 1.5894 1.2607
No log 3.4694 170 1.6774 -0.0265 1.6774 1.2952
No log 3.5102 172 2.0399 0.0392 2.0399 1.4283
No log 3.5510 174 1.9039 0.0028 1.9039 1.3798
No log 3.5918 176 1.8032 -0.0233 1.8032 1.3428
No log 3.6327 178 1.9519 0.0313 1.9519 1.3971
No log 3.6735 180 2.0668 0.0028 2.0668 1.4376
No log 3.7143 182 1.6961 0.0289 1.6961 1.3023
No log 3.7551 184 1.6565 0.0264 1.6565 1.2871
No log 3.7959 186 2.0197 -0.0221 2.0197 1.4212
No log 3.8367 188 2.0951 -0.0319 2.0951 1.4475
No log 3.8776 190 2.0192 -0.0029 2.0192 1.4210
No log 3.9184 192 1.7560 -0.0149 1.7560 1.3251
No log 3.9592 194 1.3118 0.1095 1.3118 1.1454
No log 4.0 196 1.2522 0.1292 1.2522 1.1190
No log 4.0408 198 1.4890 0.0476 1.4890 1.2203
No log 4.0816 200 1.9833 0.0437 1.9833 1.4083
No log 4.1224 202 1.9248 0.0142 1.9248 1.3874
No log 4.1633 204 1.4685 0.0405 1.4685 1.2118
No log 4.2041 206 1.0619 0.0871 1.0619 1.0305
No log 4.2449 208 1.0757 0.0871 1.0757 1.0372
No log 4.2857 210 1.4274 0.0903 1.4274 1.1948
No log 4.3265 212 1.9979 0.0457 1.9979 1.4135
No log 4.3673 214 2.0189 0.0486 2.0189 1.4209
No log 4.4082 216 1.5591 0.0 1.5591 1.2486
No log 4.4490 218 1.3694 0.0891 1.3694 1.1702
No log 4.4898 220 1.2044 0.1014 1.2044 1.0974
No log 4.5306 222 1.4125 0.0868 1.4125 1.1885
No log 4.5714 224 1.4410 0.1083 1.4410 1.2004
No log 4.6122 226 1.3380 0.1316 1.3380 1.1567
No log 4.6531 228 1.5163 0.0843 1.5163 1.2314
No log 4.6939 230 1.6385 0.0659 1.6385 1.2800
No log 4.7347 232 2.0735 0.0420 2.0735 1.4400
No log 4.7755 234 2.1033 0.0634 2.1033 1.4503
No log 4.8163 236 1.9601 0.0408 1.9601 1.4000
No log 4.8571 238 1.5605 0.0415 1.5605 1.2492
No log 4.8980 240 1.4491 0.0915 1.4491 1.2038
No log 4.9388 242 1.6610 -0.0182 1.6610 1.2888
No log 4.9796 244 1.9261 0.0866 1.9261 1.3878
No log 5.0204 246 1.7985 0.0423 1.7985 1.3411
No log 5.0612 248 1.6298 -0.0210 1.6298 1.2766
No log 5.1020 250 1.4053 0.1141 1.4053 1.1855
No log 5.1429 252 1.2628 0.0556 1.2628 1.1237
No log 5.1837 254 1.2723 0.0580 1.2723 1.1280
No log 5.2245 256 1.5799 0.0511 1.5799 1.2569
No log 5.2653 258 2.0066 0.0051 2.0066 1.4166
No log 5.3061 260 1.9358 -0.0213 1.9358 1.3913
No log 5.3469 262 1.5878 0.0523 1.5878 1.2601
No log 5.3878 264 1.2279 0.1081 1.2279 1.1081
No log 5.4286 266 1.2129 0.1608 1.2129 1.1013
No log 5.4694 268 1.1940 0.1317 1.1940 1.0927
No log 5.5102 270 1.3508 0.1373 1.3508 1.1623
No log 5.5510 272 1.5009 0.1238 1.5009 1.2251
No log 5.5918 274 1.4429 0.0933 1.4429 1.2012
No log 5.6327 276 1.4027 0.1111 1.4027 1.1843
No log 5.6735 278 1.2783 0.1081 1.2783 1.1306
No log 5.7143 280 1.3385 0.1081 1.3385 1.1569
No log 5.7551 282 1.6248 0.0991 1.6248 1.2747
No log 5.7959 284 1.8657 -0.0305 1.8657 1.3659
No log 5.8367 286 1.7597 -0.0115 1.7597 1.3265
No log 5.8776 288 1.3741 0.1068 1.3741 1.1722
No log 5.9184 290 1.1228 0.1292 1.1228 1.0596
No log 5.9592 292 1.0866 0.1292 1.0866 1.0424
No log 6.0 294 1.2407 0.0565 1.2407 1.1138
No log 6.0408 296 1.6133 0.0877 1.6133 1.2702
No log 6.0816 298 1.7661 0.0508 1.7661 1.3290
No log 6.1224 300 1.6586 -0.0029 1.6586 1.2879
No log 6.1633 302 1.4168 0.0886 1.4168 1.1903
No log 6.2041 304 1.4313 0.1083 1.4313 1.1964
No log 6.2449 306 1.6941 0.0529 1.6941 1.3016
No log 6.2857 308 2.0340 0.0549 2.0340 1.4262
No log 6.3265 310 1.9480 0.0526 1.9480 1.3957
No log 6.3673 312 1.7426 0.0241 1.7426 1.3201
No log 6.4082 314 1.5841 0.0734 1.5841 1.2586
No log 6.4490 316 1.6969 0.0276 1.6969 1.3027
No log 6.4898 318 1.6819 0.0029 1.6819 1.2969
No log 6.5306 320 1.4577 0.1041 1.4577 1.2073
No log 6.5714 322 1.3372 0.1125 1.3372 1.1564
No log 6.6122 324 1.3115 0.1125 1.3115 1.1452
No log 6.6531 326 1.4586 0.0968 1.4586 1.2077
No log 6.6939 328 1.4779 0.0927 1.4779 1.2157
No log 6.7347 330 1.5491 0.0732 1.5491 1.2446
No log 6.7755 332 1.5212 0.0927 1.5212 1.2334
No log 6.8163 334 1.3998 0.1139 1.3998 1.1831
No log 6.8571 336 1.4434 0.1139 1.4434 1.2014
No log 6.8980 338 1.4657 0.0927 1.4657 1.2107
No log 6.9388 340 1.5820 -0.0146 1.5820 1.2578
No log 6.9796 342 1.5454 0.0277 1.5454 1.2432
No log 7.0204 344 1.4450 0.0943 1.4450 1.2021
No log 7.0612 346 1.3789 0.1169 1.3789 1.1743
No log 7.1020 348 1.4249 0.0712 1.4249 1.1937
No log 7.1429 350 1.3714 0.0927 1.3714 1.1711
No log 7.1837 352 1.2954 0.1125 1.2954 1.1382
No log 7.2245 354 1.3871 0.0927 1.3871 1.1777
No log 7.2653 356 1.5052 -0.0182 1.5052 1.2269
No log 7.3061 358 1.4560 0.0943 1.4560 1.2066
No log 7.3469 360 1.4636 0.0943 1.4636 1.2098
No log 7.3878 362 1.4797 0.0464 1.4797 1.2164
No log 7.4286 364 1.3606 0.1139 1.3606 1.1664
No log 7.4694 366 1.2630 0.1111 1.2630 1.1238
No log 7.5102 368 1.2005 0.1304 1.2005 1.0957
No log 7.5510 370 1.2502 0.0831 1.2502 1.1181
No log 7.5918 372 1.4458 0.0453 1.4458 1.2024
No log 7.6327 374 1.5669 0.0296 1.5669 1.2517
No log 7.6735 376 1.5469 0.0296 1.5469 1.2438
No log 7.7143 378 1.4181 0.1111 1.4181 1.1908
No log 7.7551 380 1.2623 0.1065 1.2623 1.1235
No log 7.7959 382 1.2305 0.1329 1.2305 1.1093
No log 7.8367 384 1.2661 0.1329 1.2661 1.1252
No log 7.8776 386 1.3795 0.1111 1.3795 1.1745
No log 7.9184 388 1.5534 0.0453 1.5534 1.2464
No log 7.9592 390 1.6091 0.0541 1.6091 1.2685
No log 8.0 392 1.5977 0.0733 1.5977 1.2640
No log 8.0408 394 1.4825 0.0488 1.4825 1.2176
No log 8.0816 396 1.4326 0.0960 1.4326 1.1969
No log 8.1224 398 1.4220 0.0960 1.4220 1.1925
No log 8.1633 400 1.3671 0.0927 1.3671 1.1692
No log 8.2041 402 1.3656 0.0909 1.3656 1.1686
No log 8.2449 404 1.4283 0.0920 1.4283 1.1951
No log 8.2857 406 1.4674 0.0920 1.4674 1.2114
No log 8.3265 408 1.4589 0.1125 1.4589 1.2078
No log 8.3673 410 1.4618 0.1125 1.4618 1.2090
No log 8.4082 412 1.5226 0.1345 1.5226 1.2339
No log 8.4490 414 1.6017 0.0698 1.6017 1.2656
No log 8.4898 416 1.6176 0.0698 1.6176 1.2718
No log 8.5306 418 1.6123 0.0264 1.6123 1.2698
No log 8.5714 420 1.5896 0.0264 1.5896 1.2608
No log 8.6122 422 1.4799 0.0920 1.4799 1.2165
No log 8.6531 424 1.3486 0.1125 1.3486 1.1613
No log 8.6939 426 1.2707 0.1065 1.2707 1.1272
No log 8.7347 428 1.2695 0.0831 1.2695 1.1267
No log 8.7755 430 1.3090 0.1125 1.3090 1.1441
No log 8.8163 432 1.4018 0.0927 1.4018 1.1840
No log 8.8571 434 1.5052 0.0238 1.5052 1.2269
No log 8.8980 436 1.5559 0.0296 1.5559 1.2473
No log 8.9388 438 1.5154 0.0296 1.5154 1.2310
No log 8.9796 440 1.4278 0.0440 1.4278 1.1949
No log 9.0204 442 1.3154 0.0909 1.3154 1.1469
No log 9.0612 444 1.2374 0.1096 1.2374 1.1124
No log 9.1020 446 1.1948 0.1340 1.1948 1.0931
No log 9.1429 448 1.1979 0.1340 1.1979 1.0945
No log 9.1837 450 1.2114 0.0790 1.2114 1.1006
No log 9.2245 452 1.2643 0.1096 1.2643 1.1244
No log 9.2653 454 1.3420 0.0909 1.3420 1.1584
No log 9.3061 456 1.4405 0.0920 1.4405 1.2002
No log 9.3469 458 1.4949 0.0453 1.4949 1.2227
No log 9.3878 460 1.5012 0.0453 1.5012 1.2253
No log 9.4286 462 1.4762 0.0453 1.4762 1.2150
No log 9.4694 464 1.4348 0.0920 1.4348 1.1978
No log 9.5102 466 1.4041 0.0927 1.4041 1.1849
No log 9.5510 468 1.3790 0.1125 1.3790 1.1743
No log 9.5918 470 1.3495 0.1125 1.3495 1.1617
No log 9.6327 472 1.3418 0.1125 1.3418 1.1584
No log 9.6735 474 1.3325 0.1125 1.3325 1.1543
No log 9.7143 476 1.3190 0.1125 1.3190 1.1485
No log 9.7551 478 1.3191 0.1125 1.3191 1.1485
No log 9.7959 480 1.3209 0.1125 1.3209 1.1493
No log 9.8367 482 1.3127 0.1125 1.3127 1.1458
No log 9.8776 484 1.3087 0.1125 1.3087 1.1440
No log 9.9184 486 1.3051 0.1125 1.3051 1.1424
No log 9.9592 488 1.3009 0.1125 1.3009 1.1406
No log 10.0 490 1.2996 0.1125 1.2996 1.1400

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits4_WithDuplicationsForScore5_FineTuningAraBERT_run3_AugV5_k11_task3_organization

Finetuned
(4023)
this model