ArabicNewSplits5_FineTuningAraBERT_run1_AugV5_k10_task5_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.9280
  • Qwk: 0.6775
  • Mse: 0.9280
  • Rmse: 0.9633

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.0408 2 2.3611 0.0285 2.3611 1.5366
No log 0.0816 4 1.5354 0.2046 1.5354 1.2391
No log 0.1224 6 1.7383 -0.0443 1.7383 1.3184
No log 0.1633 8 1.8397 -0.0770 1.8397 1.3564
No log 0.2041 10 1.6591 0.0634 1.6591 1.2881
No log 0.2449 12 1.5141 0.2009 1.5141 1.2305
No log 0.2857 14 1.3729 0.2835 1.3729 1.1717
No log 0.3265 16 1.2429 0.3408 1.2429 1.1149
No log 0.3673 18 1.2527 0.2443 1.2527 1.1193
No log 0.4082 20 1.7558 0.1533 1.7558 1.3251
No log 0.4490 22 1.7247 0.1737 1.7247 1.3133
No log 0.4898 24 1.2460 0.3766 1.2460 1.1162
No log 0.5306 26 1.0726 0.3476 1.0726 1.0357
No log 0.5714 28 1.0892 0.3071 1.0892 1.0437
No log 0.6122 30 1.1129 0.3395 1.1129 1.0549
No log 0.6531 32 1.0889 0.3654 1.0889 1.0435
No log 0.6939 34 1.0790 0.3309 1.0790 1.0388
No log 0.7347 36 1.0819 0.4113 1.0819 1.0401
No log 0.7755 38 1.1150 0.4549 1.1150 1.0559
No log 0.8163 40 1.0619 0.4850 1.0619 1.0305
No log 0.8571 42 0.9755 0.4650 0.9755 0.9876
No log 0.8980 44 0.9846 0.4975 0.9846 0.9923
No log 0.9388 46 1.0198 0.4797 1.0198 1.0099
No log 0.9796 48 0.9787 0.4602 0.9787 0.9893
No log 1.0204 50 0.9365 0.4656 0.9365 0.9677
No log 1.0612 52 0.9408 0.4984 0.9408 0.9699
No log 1.1020 54 0.9210 0.5032 0.9210 0.9597
No log 1.1429 56 0.9140 0.5462 0.9140 0.9560
No log 1.1837 58 0.9231 0.5992 0.9231 0.9608
No log 1.2245 60 0.9421 0.6293 0.9421 0.9706
No log 1.2653 62 0.9682 0.5477 0.9682 0.9839
No log 1.3061 64 1.0597 0.4100 1.0597 1.0294
No log 1.3469 66 1.0953 0.3834 1.0953 1.0465
No log 1.3878 68 1.0112 0.4104 1.0112 1.0056
No log 1.4286 70 0.8936 0.5584 0.8936 0.9453
No log 1.4694 72 0.8775 0.5839 0.8775 0.9368
No log 1.5102 74 0.9205 0.5364 0.9205 0.9594
No log 1.5510 76 1.0259 0.5196 1.0259 1.0129
No log 1.5918 78 1.1194 0.5036 1.1194 1.0580
No log 1.6327 80 1.0197 0.5330 1.0197 1.0098
No log 1.6735 82 0.9431 0.5631 0.9431 0.9711
No log 1.7143 84 0.8523 0.5979 0.8523 0.9232
No log 1.7551 86 0.7975 0.5975 0.7975 0.8930
No log 1.7959 88 0.8001 0.5985 0.8001 0.8945
No log 1.8367 90 0.9486 0.6154 0.9486 0.9739
No log 1.8776 92 1.1561 0.5251 1.1561 1.0752
No log 1.9184 94 1.1594 0.5019 1.1594 1.0768
No log 1.9592 96 1.0034 0.5720 1.0034 1.0017
No log 2.0 98 0.8477 0.6363 0.8477 0.9207
No log 2.0408 100 0.8149 0.6138 0.8149 0.9027
No log 2.0816 102 0.8588 0.6320 0.8588 0.9267
No log 2.1224 104 1.0900 0.5426 1.0900 1.0440
No log 2.1633 106 1.2947 0.4657 1.2947 1.1379
No log 2.2041 108 1.2652 0.4795 1.2652 1.1248
No log 2.2449 110 1.1079 0.5791 1.1079 1.0525
No log 2.2857 112 1.0028 0.6362 1.0028 1.0014
No log 2.3265 114 0.9492 0.6290 0.9492 0.9743
No log 2.3673 116 0.9337 0.6027 0.9337 0.9663
No log 2.4082 118 1.0593 0.6043 1.0593 1.0292
No log 2.4490 120 1.3308 0.5412 1.3308 1.1536
No log 2.4898 122 1.4450 0.5579 1.4450 1.2021
No log 2.5306 124 1.4002 0.5518 1.4002 1.1833
No log 2.5714 126 1.2507 0.5623 1.2507 1.1183
No log 2.6122 128 1.0810 0.6129 1.0810 1.0397
No log 2.6531 130 0.9624 0.6807 0.9624 0.9810
No log 2.6939 132 0.9418 0.6821 0.9418 0.9705
No log 2.7347 134 0.9464 0.6965 0.9464 0.9728
No log 2.7755 136 1.1423 0.6415 1.1423 1.0688
No log 2.8163 138 1.1534 0.6543 1.1534 1.0740
No log 2.8571 140 0.8699 0.6508 0.8699 0.9327
No log 2.8980 142 0.7746 0.6828 0.7746 0.8801
No log 2.9388 144 0.8182 0.6910 0.8182 0.9045
No log 2.9796 146 0.8184 0.6968 0.8184 0.9046
No log 3.0204 148 0.8030 0.7168 0.8030 0.8961
No log 3.0612 150 0.9185 0.6821 0.9185 0.9584
No log 3.1020 152 1.1567 0.5954 1.1567 1.0755
No log 3.1429 154 1.3661 0.5690 1.3661 1.1688
No log 3.1837 156 1.2795 0.5988 1.2795 1.1312
No log 3.2245 158 1.1068 0.6405 1.1068 1.0521
No log 3.2653 160 0.9378 0.6854 0.9378 0.9684
No log 3.3061 162 0.8514 0.7009 0.8514 0.9227
No log 3.3469 164 0.9088 0.6922 0.9088 0.9533
No log 3.3878 166 1.0687 0.6450 1.0687 1.0338
No log 3.4286 168 1.3012 0.5943 1.3012 1.1407
No log 3.4694 170 1.3452 0.5907 1.3452 1.1598
No log 3.5102 172 1.2884 0.5826 1.2884 1.1351
No log 3.5510 174 1.1782 0.6189 1.1782 1.0854
No log 3.5918 176 1.0997 0.6533 1.0997 1.0487
No log 3.6327 178 1.1998 0.6361 1.1998 1.0953
No log 3.6735 180 1.2673 0.5986 1.2673 1.1257
No log 3.7143 182 1.2112 0.6023 1.2112 1.1005
No log 3.7551 184 1.1187 0.6046 1.1187 1.0577
No log 3.7959 186 1.1467 0.5724 1.1467 1.0708
No log 3.8367 188 1.0670 0.5963 1.0670 1.0329
No log 3.8776 190 0.9532 0.5966 0.9532 0.9763
No log 3.9184 192 1.0066 0.6008 1.0066 1.0033
No log 3.9592 194 1.1490 0.5789 1.1490 1.0719
No log 4.0 196 1.1985 0.6031 1.1985 1.0948
No log 4.0408 198 1.1013 0.6155 1.1013 1.0494
No log 4.0816 200 0.9703 0.6540 0.9703 0.9850
No log 4.1224 202 0.9691 0.6550 0.9691 0.9844
No log 4.1633 204 0.9960 0.6545 0.9960 0.9980
No log 4.2041 206 1.0943 0.6365 1.0943 1.0461
No log 4.2449 208 1.1548 0.6214 1.1548 1.0746
No log 4.2857 210 1.2408 0.5706 1.2408 1.1139
No log 4.3265 212 1.2335 0.5756 1.2335 1.1106
No log 4.3673 214 1.1780 0.5819 1.1780 1.0854
No log 4.4082 216 1.1653 0.6170 1.1653 1.0795
No log 4.4490 218 1.2641 0.5975 1.2641 1.1243
No log 4.4898 220 1.2287 0.6233 1.2287 1.1085
No log 4.5306 222 1.1831 0.6387 1.1831 1.0877
No log 4.5714 224 1.1030 0.6456 1.1030 1.0502
No log 4.6122 226 1.0031 0.6508 1.0031 1.0015
No log 4.6531 228 1.0286 0.6508 1.0286 1.0142
No log 4.6939 230 1.0693 0.6508 1.0693 1.0341
No log 4.7347 232 1.1133 0.6366 1.1133 1.0551
No log 4.7755 234 1.1092 0.6262 1.1092 1.0532
No log 4.8163 236 1.1685 0.6009 1.1685 1.0810
No log 4.8571 238 1.1540 0.6009 1.1540 1.0743
No log 4.8980 240 1.1505 0.6092 1.1505 1.0726
No log 4.9388 242 1.2472 0.5909 1.2472 1.1168
No log 4.9796 244 1.4047 0.5908 1.4047 1.1852
No log 5.0204 246 1.4440 0.5924 1.4440 1.2017
No log 5.0612 248 1.2476 0.6094 1.2476 1.1170
No log 5.1020 250 1.1259 0.6467 1.1259 1.0611
No log 5.1429 252 1.0639 0.6710 1.0639 1.0314
No log 5.1837 254 1.0230 0.6764 1.0230 1.0114
No log 5.2245 256 0.9740 0.6977 0.9740 0.9869
No log 5.2653 258 0.9840 0.6977 0.9840 0.9920
No log 5.3061 260 0.9372 0.6968 0.9372 0.9681
No log 5.3469 262 0.8119 0.7090 0.8119 0.9011
No log 5.3878 264 0.7745 0.7113 0.7745 0.8800
No log 5.4286 266 0.7727 0.7089 0.7727 0.8791
No log 5.4694 268 0.8876 0.7171 0.8876 0.9421
No log 5.5102 270 1.0176 0.6720 1.0176 1.0088
No log 5.5510 272 1.0466 0.6635 1.0466 1.0231
No log 5.5918 274 1.0100 0.6795 1.0100 1.0050
No log 5.6327 276 0.9334 0.7063 0.9334 0.9661
No log 5.6735 278 0.8502 0.7257 0.8502 0.9221
No log 5.7143 280 0.8568 0.7295 0.8568 0.9256
No log 5.7551 282 0.9618 0.6887 0.9618 0.9807
No log 5.7959 284 1.0965 0.6374 1.0965 1.0472
No log 5.8367 286 1.1638 0.6263 1.1638 1.0788
No log 5.8776 288 1.1588 0.6380 1.1588 1.0765
No log 5.9184 290 1.0731 0.6470 1.0731 1.0359
No log 5.9592 292 0.9824 0.6734 0.9824 0.9912
No log 6.0 294 0.9981 0.6650 0.9981 0.9991
No log 6.0408 296 0.9391 0.6792 0.9391 0.9691
No log 6.0816 298 0.8705 0.7057 0.8705 0.9330
No log 6.1224 300 0.8745 0.6883 0.8745 0.9352
No log 6.1633 302 0.9742 0.6782 0.9742 0.9870
No log 6.2041 304 1.0867 0.6620 1.0867 1.0424
No log 6.2449 306 1.1508 0.6547 1.1508 1.0727
No log 6.2857 308 1.0843 0.6620 1.0843 1.0413
No log 6.3265 310 0.9396 0.6778 0.9396 0.9693
No log 6.3673 312 0.8300 0.6879 0.8300 0.9110
No log 6.4082 314 0.8272 0.6879 0.8272 0.9095
No log 6.4490 316 0.9138 0.6679 0.9138 0.9559
No log 6.4898 318 1.0427 0.6494 1.0427 1.0211
No log 6.5306 320 1.1318 0.6364 1.1318 1.0639
No log 6.5714 322 1.1465 0.6364 1.1465 1.0707
No log 6.6122 324 1.0683 0.6597 1.0683 1.0336
No log 6.6531 326 0.9755 0.6717 0.9755 0.9877
No log 6.6939 328 0.8824 0.7082 0.8824 0.9393
No log 6.7347 330 0.8540 0.7386 0.8540 0.9241
No log 6.7755 332 0.9058 0.7288 0.9058 0.9518
No log 6.8163 334 0.9787 0.6862 0.9787 0.9893
No log 6.8571 336 1.0529 0.6877 1.0529 1.0261
No log 6.8980 338 1.0721 0.6760 1.0721 1.0354
No log 6.9388 340 1.0890 0.6623 1.0890 1.0436
No log 6.9796 342 1.0806 0.6523 1.0806 1.0395
No log 7.0204 344 1.1013 0.6501 1.1013 1.0494
No log 7.0612 346 1.0921 0.6440 1.0921 1.0451
No log 7.1020 348 1.0343 0.6418 1.0343 1.0170
No log 7.1429 350 0.9802 0.6624 0.9802 0.9901
No log 7.1837 352 0.9665 0.6548 0.9665 0.9831
No log 7.2245 354 1.0077 0.6543 1.0077 1.0038
No log 7.2653 356 1.0070 0.6543 1.0070 1.0035
No log 7.3061 358 0.9674 0.6631 0.9674 0.9836
No log 7.3469 360 0.9397 0.6721 0.9397 0.9694
No log 7.3878 362 0.9534 0.6801 0.9534 0.9764
No log 7.4286 364 1.0001 0.6562 1.0001 1.0000
No log 7.4694 366 1.0499 0.6612 1.0499 1.0246
No log 7.5102 368 1.0782 0.6441 1.0782 1.0384
No log 7.5510 370 1.0367 0.6588 1.0367 1.0182
No log 7.5918 372 1.0367 0.6588 1.0367 1.0182
No log 7.6327 374 1.0393 0.6528 1.0393 1.0195
No log 7.6735 376 1.0178 0.6528 1.0178 1.0088
No log 7.7143 378 1.0353 0.6528 1.0353 1.0175
No log 7.7551 380 0.9846 0.6498 0.9846 0.9923
No log 7.7959 382 0.9628 0.6585 0.9628 0.9812
No log 7.8367 384 0.9990 0.6393 0.9990 0.9995
No log 7.8776 386 1.0687 0.6377 1.0687 1.0338
No log 7.9184 388 1.0868 0.6377 1.0868 1.0425
No log 7.9592 390 1.0530 0.6486 1.0530 1.0262
No log 8.0 392 0.9704 0.6520 0.9704 0.9851
No log 8.0408 394 0.8890 0.7030 0.8890 0.9429
No log 8.0816 396 0.8472 0.7164 0.8472 0.9205
No log 8.1224 398 0.8448 0.7164 0.8448 0.9191
No log 8.1633 400 0.8468 0.7164 0.8468 0.9202
No log 8.2041 402 0.8901 0.7125 0.8901 0.9435
No log 8.2449 404 0.9613 0.6672 0.9613 0.9805
No log 8.2857 406 1.0312 0.6588 1.0312 1.0155
No log 8.3265 408 1.0538 0.6588 1.0538 1.0265
No log 8.3673 410 1.0546 0.6588 1.0546 1.0269
No log 8.4082 412 1.0255 0.6588 1.0255 1.0127
No log 8.4490 414 0.9993 0.6673 0.9993 0.9997
No log 8.4898 416 0.9651 0.6784 0.9651 0.9824
No log 8.5306 418 0.9219 0.6874 0.9219 0.9602
No log 8.5714 420 0.9019 0.6794 0.9019 0.9497
No log 8.6122 422 0.8968 0.6794 0.8968 0.9470
No log 8.6531 424 0.9134 0.6794 0.9134 0.9557
No log 8.6939 426 0.9336 0.6794 0.9336 0.9662
No log 8.7347 428 0.9496 0.6680 0.9496 0.9745
No log 8.7755 430 0.9927 0.6585 0.9927 0.9963
No log 8.8163 432 1.0276 0.6302 1.0276 1.0137
No log 8.8571 434 1.0753 0.6396 1.0753 1.0370
No log 8.8980 436 1.0837 0.6396 1.0837 1.0410
No log 8.9388 438 1.0609 0.6396 1.0609 1.0300
No log 8.9796 440 1.0195 0.6427 1.0195 1.0097
No log 9.0204 442 0.9777 0.6585 0.9777 0.9888
No log 9.0612 444 0.9549 0.6761 0.9549 0.9772
No log 9.1020 446 0.9378 0.6874 0.9378 0.9684
No log 9.1429 448 0.9364 0.6874 0.9364 0.9677
No log 9.1837 450 0.9403 0.6864 0.9403 0.9697
No log 9.2245 452 0.9448 0.6864 0.9448 0.9720
No log 9.2653 454 0.9443 0.6864 0.9443 0.9718
No log 9.3061 456 0.9322 0.6864 0.9322 0.9655
No log 9.3469 458 0.9221 0.6864 0.9221 0.9603
No log 9.3878 460 0.9265 0.6864 0.9265 0.9626
No log 9.4286 462 0.9306 0.6864 0.9306 0.9647
No log 9.4694 464 0.9359 0.6775 0.9359 0.9674
No log 9.5102 466 0.9523 0.6853 0.9523 0.9758
No log 9.5510 468 0.9617 0.6853 0.9617 0.9807
No log 9.5918 470 0.9686 0.6752 0.9686 0.9842
No log 9.6327 472 0.9745 0.6752 0.9745 0.9872
No log 9.6735 474 0.9725 0.6752 0.9725 0.9862
No log 9.7143 476 0.9644 0.6752 0.9644 0.9821
No log 9.7551 478 0.9588 0.6743 0.9588 0.9792
No log 9.7959 480 0.9552 0.6853 0.9552 0.9773
No log 9.8367 482 0.9481 0.6853 0.9481 0.9737
No log 9.8776 484 0.9404 0.6775 0.9404 0.9697
No log 9.9184 486 0.9332 0.6775 0.9332 0.9660
No log 9.9592 488 0.9291 0.6775 0.9291 0.9639
No log 10.0 490 0.9280 0.6775 0.9280 0.9633

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits5_FineTuningAraBERT_run1_AugV5_k10_task5_organization

Finetuned
(4023)
this model