ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k13_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5017
  • Qwk: 0.5591
  • Mse: 0.5017
  • Rmse: 0.7083

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0455 2 2.5957 -0.0230 2.5957 1.6111
No log 0.0909 4 1.3124 0.0745 1.3124 1.1456
No log 0.1364 6 0.9584 -0.0228 0.9584 0.9790
No log 0.1818 8 0.7845 0.0481 0.7845 0.8857
No log 0.2273 10 0.7016 0.0481 0.7016 0.8376
No log 0.2727 12 0.7205 0.1365 0.7205 0.8488
No log 0.3182 14 0.8193 0.2736 0.8193 0.9052
No log 0.3636 16 0.7242 0.2156 0.7242 0.8510
No log 0.4091 18 0.8111 0.3598 0.8111 0.9006
No log 0.4545 20 0.7386 0.3060 0.7386 0.8594
No log 0.5 22 0.6565 0.2808 0.6565 0.8103
No log 0.5455 24 0.6403 0.2206 0.6403 0.8002
No log 0.5909 26 0.7721 0.1007 0.7721 0.8787
No log 0.6364 28 0.7882 0.0 0.7882 0.8878
No log 0.6818 30 0.8345 0.0053 0.8345 0.9135
No log 0.7273 32 0.8237 0.0947 0.8237 0.9076
No log 0.7727 34 0.8258 0.2109 0.8258 0.9087
No log 0.8182 36 0.7491 0.1739 0.7491 0.8655
No log 0.8636 38 0.6643 0.1770 0.6643 0.8151
No log 0.9091 40 0.5996 0.1184 0.5996 0.7743
No log 0.9545 42 0.5930 0.3029 0.5930 0.7701
No log 1.0 44 0.5896 0.2955 0.5896 0.7678
No log 1.0455 46 0.6528 0.3071 0.6528 0.8080
No log 1.0909 48 0.8077 0.3094 0.8077 0.8987
No log 1.1364 50 0.7273 0.3417 0.7273 0.8528
No log 1.1818 52 0.6374 0.4083 0.6374 0.7984
No log 1.2273 54 0.6087 0.4322 0.6087 0.7802
No log 1.2727 56 0.7073 0.4258 0.7073 0.8410
No log 1.3182 58 0.7667 0.5146 0.7667 0.8756
No log 1.3636 60 0.6246 0.4351 0.6246 0.7903
No log 1.4091 62 0.5295 0.4857 0.5295 0.7277
No log 1.4545 64 0.5861 0.4260 0.5861 0.7655
No log 1.5 66 0.6499 0.4777 0.6499 0.8062
No log 1.5455 68 0.5800 0.4924 0.5800 0.7616
No log 1.5909 70 0.5838 0.5554 0.5838 0.7641
No log 1.6364 72 0.6375 0.5384 0.6375 0.7984
No log 1.6818 74 0.6249 0.3829 0.6249 0.7905
No log 1.7273 76 0.6232 0.2890 0.6232 0.7894
No log 1.7727 78 0.7433 0.3329 0.7433 0.8622
No log 1.8182 80 0.7184 0.2913 0.7184 0.8476
No log 1.8636 82 0.6707 0.1550 0.6707 0.8190
No log 1.9091 84 0.8313 0.1334 0.8313 0.9117
No log 1.9545 86 1.0809 0.3410 1.0809 1.0396
No log 2.0 88 1.0699 0.3760 1.0699 1.0344
No log 2.0455 90 0.8719 0.2942 0.8719 0.9337
No log 2.0909 92 0.7149 0.0481 0.7149 0.8455
No log 2.1364 94 0.6435 0.1508 0.6435 0.8022
No log 2.1818 96 0.6627 0.2817 0.6627 0.8141
No log 2.2273 98 0.6754 0.3637 0.6754 0.8218
No log 2.2727 100 0.6021 0.2285 0.6021 0.7759
No log 2.3182 102 0.6326 0.2041 0.6326 0.7954
No log 2.3636 104 0.7848 0.3230 0.7848 0.8859
No log 2.4091 106 0.8520 0.2958 0.8520 0.9230
No log 2.4545 108 0.7658 0.3697 0.7658 0.8751
No log 2.5 110 0.5996 0.4249 0.5996 0.7744
No log 2.5455 112 0.5612 0.4829 0.5612 0.7491
No log 2.5909 114 0.6392 0.4193 0.6392 0.7995
No log 2.6364 116 0.7990 0.3663 0.7990 0.8939
No log 2.6818 118 0.8609 0.3643 0.8609 0.9279
No log 2.7273 120 0.7652 0.3051 0.7652 0.8748
No log 2.7727 122 0.6118 0.3477 0.6118 0.7822
No log 2.8182 124 0.5552 0.3745 0.5552 0.7451
No log 2.8636 126 0.5326 0.3608 0.5326 0.7298
No log 2.9091 128 0.5432 0.3729 0.5432 0.7370
No log 2.9545 130 0.6733 0.4725 0.6733 0.8206
No log 3.0 132 0.7794 0.4195 0.7794 0.8828
No log 3.0455 134 0.6956 0.4921 0.6956 0.8340
No log 3.0909 136 0.5804 0.4825 0.5804 0.7618
No log 3.1364 138 0.5522 0.5230 0.5522 0.7431
No log 3.1818 140 0.5454 0.5422 0.5454 0.7385
No log 3.2273 142 0.5285 0.5611 0.5285 0.7270
No log 3.2727 144 0.5168 0.6183 0.5168 0.7189
No log 3.3182 146 0.5215 0.6198 0.5215 0.7222
No log 3.3636 148 0.5328 0.5184 0.5328 0.7299
No log 3.4091 150 0.5183 0.6222 0.5183 0.7199
No log 3.4545 152 0.5045 0.6267 0.5045 0.7103
No log 3.5 154 0.5058 0.6053 0.5058 0.7112
No log 3.5455 156 0.5639 0.5153 0.5639 0.7509
No log 3.5909 158 0.6026 0.4395 0.6026 0.7763
No log 3.6364 160 0.5294 0.4869 0.5294 0.7276
No log 3.6818 162 0.5214 0.4762 0.5214 0.7221
No log 3.7273 164 0.5146 0.5463 0.5146 0.7174
No log 3.7727 166 0.5266 0.5195 0.5266 0.7257
No log 3.8182 168 0.5402 0.5569 0.5402 0.7350
No log 3.8636 170 0.5301 0.5781 0.5301 0.7280
No log 3.9091 172 0.5447 0.5553 0.5447 0.7380
No log 3.9545 174 0.5885 0.5765 0.5885 0.7671
No log 4.0 176 0.6844 0.4719 0.6844 0.8273
No log 4.0455 178 0.5840 0.5436 0.5840 0.7642
No log 4.0909 180 0.5126 0.6034 0.5126 0.7160
No log 4.1364 182 0.5321 0.5273 0.5321 0.7295
No log 4.1818 184 0.5584 0.4933 0.5584 0.7473
No log 4.2273 186 0.5051 0.5190 0.5051 0.7107
No log 4.2727 188 0.6648 0.4687 0.6648 0.8153
No log 4.3182 190 0.7883 0.4133 0.7883 0.8879
No log 4.3636 192 0.6912 0.4161 0.6912 0.8314
No log 4.4091 194 0.5340 0.5904 0.5340 0.7308
No log 4.4545 196 0.5864 0.4218 0.5864 0.7658
No log 4.5 198 0.5953 0.4330 0.5953 0.7716
No log 4.5455 200 0.5545 0.4105 0.5545 0.7447
No log 4.5909 202 0.6574 0.4246 0.6574 0.8108
No log 4.6364 204 0.7022 0.3620 0.7022 0.8380
No log 4.6818 206 0.6200 0.3878 0.6200 0.7874
No log 4.7273 208 0.5426 0.4590 0.5426 0.7366
No log 4.7727 210 0.5296 0.5003 0.5296 0.7278
No log 4.8182 212 0.5182 0.4743 0.5182 0.7199
No log 4.8636 214 0.4931 0.5888 0.4931 0.7022
No log 4.9091 216 0.5525 0.5721 0.5525 0.7433
No log 4.9545 218 0.6094 0.5328 0.6094 0.7806
No log 5.0 220 0.5436 0.6104 0.5436 0.7373
No log 5.0455 222 0.4829 0.6326 0.4829 0.6949
No log 5.0909 224 0.4827 0.6796 0.4827 0.6947
No log 5.1364 226 0.4760 0.6494 0.4760 0.6899
No log 5.1818 228 0.5104 0.6104 0.5104 0.7144
No log 5.2273 230 0.4869 0.6427 0.4869 0.6977
No log 5.2727 232 0.4358 0.6747 0.4358 0.6602
No log 5.3182 234 0.4530 0.6201 0.4530 0.6731
No log 5.3636 236 0.4881 0.5560 0.4881 0.6987
No log 5.4091 238 0.4400 0.6339 0.4400 0.6633
No log 5.4545 240 0.4515 0.7245 0.4515 0.6719
No log 5.5 242 0.4874 0.6248 0.4874 0.6982
No log 5.5455 244 0.4530 0.7236 0.4530 0.6731
No log 5.5909 246 0.4564 0.5912 0.4564 0.6756
No log 5.6364 248 0.4906 0.5166 0.4906 0.7004
No log 5.6818 250 0.4674 0.5786 0.4674 0.6837
No log 5.7273 252 0.4608 0.5937 0.4608 0.6788
No log 5.7727 254 0.4560 0.6032 0.4560 0.6753
No log 5.8182 256 0.4731 0.5266 0.4731 0.6878
No log 5.8636 258 0.5184 0.5368 0.5184 0.7200
No log 5.9091 260 0.5204 0.5161 0.5204 0.7214
No log 5.9545 262 0.4641 0.6335 0.4641 0.6812
No log 6.0 264 0.4527 0.6184 0.4527 0.6728
No log 6.0455 266 0.4482 0.6060 0.4482 0.6695
No log 6.0909 268 0.4412 0.6269 0.4412 0.6642
No log 6.1364 270 0.5048 0.5794 0.5048 0.7105
No log 6.1818 272 0.6068 0.4878 0.6068 0.7790
No log 6.2273 274 0.5934 0.5228 0.5934 0.7703
No log 6.2727 276 0.4963 0.5195 0.4963 0.7045
No log 6.3182 278 0.4426 0.6439 0.4426 0.6653
No log 6.3636 280 0.4611 0.5184 0.4611 0.6791
No log 6.4091 282 0.4738 0.5184 0.4738 0.6884
No log 6.4545 284 0.4732 0.5272 0.4732 0.6879
No log 6.5 286 0.4608 0.6214 0.4608 0.6788
No log 6.5455 288 0.4628 0.6403 0.4628 0.6803
No log 6.5909 290 0.4796 0.5569 0.4796 0.6926
No log 6.6364 292 0.4810 0.5569 0.4810 0.6936
No log 6.6818 294 0.4629 0.6229 0.4629 0.6804
No log 6.7273 296 0.5031 0.6514 0.5031 0.7093
No log 6.7727 298 0.5784 0.4987 0.5784 0.7605
No log 6.8182 300 0.5471 0.5358 0.5471 0.7397
No log 6.8636 302 0.4775 0.6928 0.4775 0.6910
No log 6.9091 304 0.4733 0.6021 0.4733 0.6880
No log 6.9545 306 0.4646 0.5974 0.4646 0.6816
No log 7.0 308 0.4444 0.6530 0.4444 0.6666
No log 7.0455 310 0.4695 0.6317 0.4695 0.6852
No log 7.0909 312 0.4929 0.6127 0.4929 0.7021
No log 7.1364 314 0.4707 0.6210 0.4707 0.6860
No log 7.1818 316 0.4374 0.6967 0.4374 0.6614
No log 7.2273 318 0.4549 0.6032 0.4549 0.6745
No log 7.2727 320 0.4536 0.6032 0.4536 0.6735
No log 7.3182 322 0.4592 0.6158 0.4592 0.6776
No log 7.3636 324 0.4711 0.5816 0.4711 0.6863
No log 7.4091 326 0.4804 0.5671 0.4804 0.6931
No log 7.4545 328 0.4816 0.5003 0.4816 0.6940
No log 7.5 330 0.4944 0.5076 0.4944 0.7032
No log 7.5455 332 0.5016 0.4962 0.5016 0.7082
No log 7.5909 334 0.5093 0.5448 0.5093 0.7137
No log 7.6364 336 0.4776 0.4888 0.4776 0.6911
No log 7.6818 338 0.4608 0.5081 0.4608 0.6788
No log 7.7273 340 0.4491 0.6154 0.4491 0.6701
No log 7.7727 342 0.4465 0.6759 0.4465 0.6682
No log 7.8182 344 0.4636 0.6601 0.4636 0.6808
No log 7.8636 346 0.5139 0.5657 0.5139 0.7169
No log 7.9091 348 0.5600 0.5161 0.5600 0.7483
No log 7.9545 350 0.5571 0.5161 0.5571 0.7464
No log 8.0 352 0.5142 0.5356 0.5142 0.7171
No log 8.0455 354 0.5147 0.4981 0.5147 0.7175
No log 8.0909 356 0.5267 0.5189 0.5267 0.7257
No log 8.1364 358 0.5104 0.5430 0.5104 0.7144
No log 8.1818 360 0.4844 0.6242 0.4844 0.6960
No log 8.2273 362 0.5284 0.5569 0.5284 0.7269
No log 8.2727 364 0.5259 0.5779 0.5259 0.7252
No log 8.3182 366 0.5278 0.5584 0.5278 0.7265
No log 8.3636 368 0.5224 0.5733 0.5224 0.7228
No log 8.4091 370 0.5248 0.5244 0.5248 0.7244
No log 8.4545 372 0.5442 0.5244 0.5442 0.7377
No log 8.5 374 0.5605 0.5189 0.5605 0.7487
No log 8.5455 376 0.5602 0.5334 0.5602 0.7484
No log 8.5909 378 0.5238 0.5657 0.5238 0.7237
No log 8.6364 380 0.5110 0.5983 0.5110 0.7148
No log 8.6818 382 0.5606 0.5552 0.5606 0.7487
No log 8.7273 384 0.5431 0.5945 0.5431 0.7369
No log 8.7727 386 0.4779 0.6130 0.4779 0.6913
No log 8.8182 388 0.4491 0.6158 0.4491 0.6701
No log 8.8636 390 0.4569 0.6018 0.4569 0.6759
No log 8.9091 392 0.4601 0.5289 0.4601 0.6783
No log 8.9545 394 0.4483 0.6010 0.4483 0.6695
No log 9.0 396 0.4507 0.5815 0.4507 0.6713
No log 9.0455 398 0.4809 0.5178 0.4809 0.6935
No log 9.0909 400 0.4867 0.5110 0.4867 0.6976
No log 9.1364 402 0.4593 0.6087 0.4593 0.6777
No log 9.1818 404 0.4562 0.5725 0.4562 0.6754
No log 9.2273 406 0.4615 0.5586 0.4615 0.6794
No log 9.2727 408 0.5081 0.5848 0.5081 0.7128
No log 9.3182 410 0.6538 0.4601 0.6538 0.8086
No log 9.3636 412 0.7210 0.4321 0.7210 0.8491
No log 9.4091 414 0.6326 0.4961 0.6326 0.7954
No log 9.4545 416 0.4944 0.5918 0.4944 0.7031
No log 9.5 418 0.4516 0.5749 0.4516 0.6720
No log 9.5455 420 0.4977 0.5468 0.4977 0.7055
No log 9.5909 422 0.4897 0.5560 0.4897 0.6998
No log 9.6364 424 0.4496 0.6228 0.4496 0.6705
No log 9.6818 426 0.4322 0.6541 0.4322 0.6574
No log 9.7273 428 0.4421 0.5867 0.4421 0.6649
No log 9.7727 430 0.4411 0.5867 0.4411 0.6641
No log 9.8182 432 0.4569 0.6365 0.4569 0.6759
No log 9.8636 434 0.4902 0.6344 0.4902 0.7001
No log 9.9091 436 0.4621 0.6348 0.4621 0.6798
No log 9.9545 438 0.4338 0.5953 0.4338 0.6586
No log 10.0 440 0.4641 0.6127 0.4641 0.6813
No log 10.0455 442 0.4700 0.6127 0.4700 0.6856
No log 10.0909 444 0.4328 0.5999 0.4328 0.6579
No log 10.1364 446 0.3965 0.6942 0.3965 0.6297
No log 10.1818 448 0.3958 0.7044 0.3958 0.6291
No log 10.2273 450 0.3960 0.7248 0.3960 0.6293
No log 10.2727 452 0.3981 0.7154 0.3981 0.6310
No log 10.3182 454 0.4330 0.6210 0.4330 0.6580
No log 10.3636 456 0.5045 0.6181 0.5045 0.7103
No log 10.4091 458 0.5005 0.6181 0.5005 0.7074
No log 10.4545 460 0.4383 0.6210 0.4383 0.6620
No log 10.5 462 0.4086 0.7133 0.4086 0.6392
No log 10.5455 464 0.4136 0.6931 0.4136 0.6432
No log 10.5909 466 0.4117 0.6712 0.4117 0.6416
No log 10.6364 468 0.4013 0.7217 0.4013 0.6335
No log 10.6818 470 0.4263 0.6295 0.4263 0.6529
No log 10.7273 472 0.4662 0.6330 0.4662 0.6828
No log 10.7727 474 0.5013 0.5983 0.5013 0.7080
No log 10.8182 476 0.4621 0.6442 0.4621 0.6798
No log 10.8636 478 0.4044 0.5999 0.4044 0.6359
No log 10.9091 480 0.3887 0.6477 0.3887 0.6235
No log 10.9545 482 0.3859 0.6477 0.3859 0.6212
No log 11.0 484 0.4138 0.6210 0.4138 0.6433
No log 11.0455 486 0.4999 0.5922 0.4999 0.7070
No log 11.0909 488 0.5229 0.5735 0.5229 0.7232
No log 11.1364 490 0.4740 0.5657 0.4740 0.6885
No log 11.1818 492 0.4084 0.6082 0.4084 0.6391
No log 11.2273 494 0.3836 0.7161 0.3836 0.6193
No log 11.2727 496 0.4003 0.6517 0.4003 0.6327
No log 11.3182 498 0.4085 0.7154 0.4085 0.6391
0.3236 11.3636 500 0.4101 0.6566 0.4101 0.6404
0.3236 11.4091 502 0.4380 0.6101 0.4380 0.6618
0.3236 11.4545 504 0.4395 0.6491 0.4395 0.6629
0.3236 11.5 506 0.4148 0.6479 0.4148 0.6440
0.3236 11.5455 508 0.4540 0.6271 0.4540 0.6738
0.3236 11.5909 510 0.5017 0.5591 0.5017 0.7083

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k13_task7_organization

Finetuned
(4019)
this model