ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k4_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.5103
  • Qwk: 0.6101
  • Mse: 0.5103
  • Rmse: 0.7144

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.125 2 2.5751 -0.0262 2.5751 1.6047
No log 0.25 4 1.2794 0.0750 1.2794 1.1311
No log 0.375 6 1.0093 -0.0970 1.0093 1.0046
No log 0.5 8 1.0537 -0.1580 1.0537 1.0265
No log 0.625 10 0.9062 0.0428 0.9062 0.9520
No log 0.75 12 0.7158 -0.0054 0.7158 0.8460
No log 0.875 14 0.7007 0.0444 0.7007 0.8371
No log 1.0 16 0.7387 0.2526 0.7387 0.8595
No log 1.125 18 0.8972 0.2908 0.8972 0.9472
No log 1.25 20 0.8091 0.2356 0.8091 0.8995
No log 1.375 22 0.6622 0.2103 0.6622 0.8137
No log 1.5 24 0.8248 0.2409 0.8248 0.9082
No log 1.625 26 0.8994 0.1672 0.8994 0.9484
No log 1.75 28 0.7970 0.1407 0.7970 0.8928
No log 1.875 30 0.7009 0.1372 0.7009 0.8372
No log 2.0 32 0.7988 0.2464 0.7988 0.8937
No log 2.125 34 0.9361 0.3085 0.9361 0.9675
No log 2.25 36 0.9278 0.3155 0.9278 0.9632
No log 2.375 38 0.8371 0.2558 0.8371 0.9149
No log 2.5 40 0.7596 0.0 0.7596 0.8715
No log 2.625 42 0.8100 0.1050 0.8100 0.9000
No log 2.75 44 0.9418 0.0541 0.9418 0.9705
No log 2.875 46 0.9637 0.1313 0.9637 0.9817
No log 3.0 48 0.8989 0.1313 0.8989 0.9481
No log 3.125 50 0.8362 0.1699 0.8362 0.9144
No log 3.25 52 0.7700 0.0757 0.7700 0.8775
No log 3.375 54 0.7355 0.0893 0.7355 0.8576
No log 3.5 56 0.7328 0.1770 0.7328 0.8560
No log 3.625 58 0.7802 0.3799 0.7802 0.8833
No log 3.75 60 0.6729 0.3772 0.6729 0.8203
No log 3.875 62 0.6324 0.1529 0.6324 0.7952
No log 4.0 64 0.5978 0.3183 0.5978 0.7732
No log 4.125 66 0.5452 0.4878 0.5452 0.7384
No log 4.25 68 0.6092 0.4249 0.6092 0.7805
No log 4.375 70 0.7313 0.2891 0.7313 0.8551
No log 4.5 72 0.6978 0.2299 0.6978 0.8353
No log 4.625 74 0.6151 0.3868 0.6151 0.7843
No log 4.75 76 0.6690 0.3972 0.6690 0.8179
No log 4.875 78 0.6671 0.4424 0.6671 0.8168
No log 5.0 80 0.6343 0.4134 0.6343 0.7965
No log 5.125 82 0.5709 0.4378 0.5709 0.7556
No log 5.25 84 0.5823 0.5227 0.5823 0.7631
No log 5.375 86 0.6142 0.4345 0.6142 0.7837
No log 5.5 88 0.5926 0.4990 0.5926 0.7698
No log 5.625 90 0.5787 0.4681 0.5787 0.7608
No log 5.75 92 0.5706 0.5681 0.5706 0.7554
No log 5.875 94 0.5976 0.5750 0.5976 0.7731
No log 6.0 96 0.8231 0.3128 0.8231 0.9073
No log 6.125 98 1.1547 0.2477 1.1547 1.0746
No log 6.25 100 0.9590 0.2581 0.9590 0.9793
No log 6.375 102 0.6002 0.5611 0.6002 0.7747
No log 6.5 104 0.6951 0.3844 0.6951 0.8337
No log 6.625 106 0.7598 0.4222 0.7598 0.8717
No log 6.75 108 0.5907 0.5201 0.5907 0.7686
No log 6.875 110 0.5848 0.5979 0.5848 0.7647
No log 7.0 112 0.6749 0.3912 0.6749 0.8215
No log 7.125 114 0.7198 0.4063 0.7198 0.8484
No log 7.25 116 0.6267 0.5589 0.6267 0.7916
No log 7.375 118 0.5613 0.5648 0.5613 0.7492
No log 7.5 120 0.6782 0.4436 0.6782 0.8235
No log 7.625 122 0.7845 0.4604 0.7845 0.8857
No log 7.75 124 0.6813 0.4799 0.6813 0.8254
No log 7.875 126 0.5122 0.5852 0.5122 0.7157
No log 8.0 128 0.5258 0.5687 0.5258 0.7251
No log 8.125 130 0.5644 0.5406 0.5644 0.7513
No log 8.25 132 0.5556 0.4127 0.5556 0.7454
No log 8.375 134 0.5521 0.5056 0.5521 0.7430
No log 8.5 136 0.6759 0.5107 0.6759 0.8221
No log 8.625 138 0.7321 0.4592 0.7321 0.8556
No log 8.75 140 0.6409 0.5342 0.6409 0.8006
No log 8.875 142 0.5784 0.5437 0.5784 0.7606
No log 9.0 144 0.5700 0.5195 0.5700 0.7550
No log 9.125 146 0.6382 0.3920 0.6382 0.7989
No log 9.25 148 0.6513 0.4160 0.6513 0.8070
No log 9.375 150 0.5909 0.4636 0.5909 0.7687
No log 9.5 152 0.5498 0.5640 0.5498 0.7415
No log 9.625 154 0.5782 0.5093 0.5782 0.7604
No log 9.75 156 0.6123 0.5246 0.6123 0.7825
No log 9.875 158 0.5681 0.5272 0.5681 0.7537
No log 10.0 160 0.5292 0.5941 0.5292 0.7275
No log 10.125 162 0.5439 0.5373 0.5439 0.7375
No log 10.25 164 0.5307 0.5373 0.5307 0.7285
No log 10.375 166 0.5424 0.5151 0.5424 0.7365
No log 10.5 168 0.5910 0.5624 0.5910 0.7687
No log 10.625 170 0.6175 0.5096 0.6175 0.7858
No log 10.75 172 0.5725 0.5617 0.5725 0.7566
No log 10.875 174 0.5358 0.5941 0.5358 0.7320
No log 11.0 176 0.5309 0.6039 0.5309 0.7286
No log 11.125 178 0.5358 0.5719 0.5358 0.7320
No log 11.25 180 0.5521 0.6326 0.5521 0.7430
No log 11.375 182 0.5520 0.6065 0.5520 0.7430
No log 11.5 184 0.5447 0.6001 0.5447 0.7380
No log 11.625 186 0.5349 0.5875 0.5349 0.7314
No log 11.75 188 0.5273 0.5875 0.5273 0.7261
No log 11.875 190 0.5238 0.5875 0.5238 0.7237
No log 12.0 192 0.5229 0.5875 0.5229 0.7231
No log 12.125 194 0.5286 0.5875 0.5286 0.7270
No log 12.25 196 0.5366 0.5875 0.5366 0.7326
No log 12.375 198 0.5541 0.4693 0.5541 0.7444
No log 12.5 200 0.5742 0.4636 0.5742 0.7577
No log 12.625 202 0.5493 0.5390 0.5493 0.7411
No log 12.75 204 0.5453 0.5567 0.5453 0.7384
No log 12.875 206 0.6143 0.4864 0.6143 0.7838
No log 13.0 208 0.6299 0.5034 0.6299 0.7937
No log 13.125 210 0.6242 0.5184 0.6242 0.7901
No log 13.25 212 0.5765 0.5648 0.5765 0.7593
No log 13.375 214 0.5675 0.4738 0.5675 0.7533
No log 13.5 216 0.5611 0.5195 0.5611 0.7491
No log 13.625 218 0.5531 0.5177 0.5531 0.7437
No log 13.75 220 0.5463 0.5177 0.5463 0.7391
No log 13.875 222 0.5543 0.6038 0.5543 0.7445
No log 14.0 224 0.6160 0.5605 0.6160 0.7848
No log 14.125 226 0.7058 0.5076 0.7058 0.8401
No log 14.25 228 0.6602 0.5336 0.6602 0.8125
No log 14.375 230 0.5295 0.6028 0.5295 0.7277
No log 14.5 232 0.4986 0.6133 0.4986 0.7061
No log 14.625 234 0.4973 0.6133 0.4973 0.7052
No log 14.75 236 0.5055 0.6233 0.5055 0.7110
No log 14.875 238 0.5709 0.5605 0.5709 0.7556
No log 15.0 240 0.5446 0.5239 0.5446 0.7380
No log 15.125 242 0.5031 0.5965 0.5031 0.7093
No log 15.25 244 0.5090 0.6092 0.5090 0.7134
No log 15.375 246 0.5584 0.5123 0.5584 0.7473
No log 15.5 248 0.5987 0.5140 0.5987 0.7738
No log 15.625 250 0.5836 0.5817 0.5836 0.7640
No log 15.75 252 0.5394 0.5307 0.5394 0.7344
No log 15.875 254 0.4994 0.5719 0.4994 0.7067
No log 16.0 256 0.4852 0.5941 0.4852 0.6965
No log 16.125 258 0.4816 0.6039 0.4816 0.6940
No log 16.25 260 0.4812 0.6039 0.4812 0.6937
No log 16.375 262 0.4957 0.5648 0.4957 0.7041
No log 16.5 264 0.5000 0.5648 0.5000 0.7071
No log 16.625 266 0.5011 0.5749 0.5011 0.7079
No log 16.75 268 0.5126 0.5765 0.5126 0.7160
No log 16.875 270 0.5301 0.5475 0.5301 0.7281
No log 17.0 272 0.5451 0.5633 0.5451 0.7383
No log 17.125 274 0.5398 0.5623 0.5398 0.7347
No log 17.25 276 0.5394 0.5853 0.5394 0.7345
No log 17.375 278 0.5447 0.5404 0.5447 0.7381
No log 17.5 280 0.5452 0.5404 0.5452 0.7384
No log 17.625 282 0.5353 0.5853 0.5353 0.7316
No log 17.75 284 0.5703 0.4575 0.5703 0.7552
No log 17.875 286 0.5633 0.4850 0.5633 0.7506
No log 18.0 288 0.5265 0.6053 0.5265 0.7256
No log 18.125 290 0.5173 0.6091 0.5173 0.7192
No log 18.25 292 0.5454 0.5852 0.5454 0.7385
No log 18.375 294 0.5787 0.5544 0.5787 0.7607
No log 18.5 296 0.5424 0.5852 0.5424 0.7365
No log 18.625 298 0.5222 0.5765 0.5222 0.7226
No log 18.75 300 0.5244 0.5765 0.5244 0.7242
No log 18.875 302 0.5409 0.5605 0.5409 0.7354
No log 19.0 304 0.5666 0.4983 0.5666 0.7527
No log 19.125 306 0.5675 0.4983 0.5675 0.7533
No log 19.25 308 0.5579 0.5248 0.5579 0.7469
No log 19.375 310 0.5355 0.5853 0.5355 0.7318
No log 19.5 312 0.5243 0.5782 0.5243 0.7241
No log 19.625 314 0.5093 0.5734 0.5093 0.7137
No log 19.75 316 0.5110 0.5883 0.5110 0.7149
No log 19.875 318 0.5249 0.6096 0.5249 0.7245
No log 20.0 320 0.5194 0.6096 0.5194 0.7207
No log 20.125 322 0.5196 0.5868 0.5196 0.7208
No log 20.25 324 0.5497 0.5614 0.5497 0.7414
No log 20.375 326 0.5432 0.5614 0.5432 0.7370
No log 20.5 328 0.5259 0.5631 0.5259 0.7252
No log 20.625 330 0.5495 0.5614 0.5495 0.7413
No log 20.75 332 0.5629 0.5420 0.5629 0.7503
No log 20.875 334 0.5556 0.5195 0.5556 0.7454
No log 21.0 336 0.5885 0.4845 0.5885 0.7671
No log 21.125 338 0.6070 0.5086 0.6070 0.7791
No log 21.25 340 0.5668 0.4660 0.5668 0.7528
No log 21.375 342 0.5566 0.4657 0.5566 0.7461
No log 21.5 344 0.6092 0.4813 0.6092 0.7805
No log 21.625 346 0.6050 0.4461 0.6050 0.7778
No log 21.75 348 0.5669 0.5009 0.5669 0.7530
No log 21.875 350 0.5438 0.5656 0.5438 0.7374
No log 22.0 352 0.5518 0.5567 0.5518 0.7428
No log 22.125 354 0.5546 0.6018 0.5546 0.7447
No log 22.25 356 0.5447 0.5937 0.5447 0.7380
No log 22.375 358 0.5331 0.6129 0.5331 0.7301
No log 22.5 360 0.5321 0.6087 0.5321 0.7295
No log 22.625 362 0.5610 0.4340 0.5610 0.7490
No log 22.75 364 0.5831 0.4410 0.5831 0.7636
No log 22.875 366 0.5551 0.5212 0.5551 0.7451
No log 23.0 368 0.5347 0.5503 0.5347 0.7312
No log 23.125 370 0.5586 0.4888 0.5586 0.7474
No log 23.25 372 0.5731 0.4330 0.5731 0.7570
No log 23.375 374 0.5628 0.5151 0.5628 0.7502
No log 23.5 376 0.5638 0.5042 0.5638 0.7508
No log 23.625 378 0.5728 0.4991 0.5728 0.7568
No log 23.75 380 0.5801 0.4386 0.5801 0.7616
No log 23.875 382 0.5737 0.5057 0.5737 0.7574
No log 24.0 384 0.5573 0.5389 0.5573 0.7465
No log 24.125 386 0.5793 0.6341 0.5793 0.7611
No log 24.25 388 0.5646 0.6334 0.5646 0.7514
No log 24.375 390 0.5481 0.6032 0.5481 0.7403
No log 24.5 392 0.5385 0.5640 0.5385 0.7338
No log 24.625 394 0.5386 0.6111 0.5386 0.7339
No log 24.75 396 0.5463 0.5912 0.5463 0.7391
No log 24.875 398 0.5845 0.5157 0.5845 0.7645
No log 25.0 400 0.6186 0.5059 0.6186 0.7865
No log 25.125 402 0.6482 0.4486 0.6482 0.8051
No log 25.25 404 0.6437 0.4845 0.6437 0.8023
No log 25.375 406 0.6012 0.5512 0.6012 0.7754
No log 25.5 408 0.5624 0.5455 0.5624 0.7499
No log 25.625 410 0.5485 0.5455 0.5485 0.7406
No log 25.75 412 0.5351 0.5719 0.5351 0.7315
No log 25.875 414 0.5281 0.6014 0.5281 0.7267
No log 26.0 416 0.5237 0.6257 0.5237 0.7237
No log 26.125 418 0.5240 0.6183 0.5240 0.7239
No log 26.25 420 0.5207 0.6173 0.5207 0.7216
No log 26.375 422 0.5251 0.5899 0.5251 0.7247
No log 26.5 424 0.5165 0.5874 0.5165 0.7187
No log 26.625 426 0.5149 0.5951 0.5149 0.7175
No log 26.75 428 0.5211 0.5593 0.5211 0.7219
No log 26.875 430 0.5212 0.5567 0.5212 0.7220
No log 27.0 432 0.5372 0.5160 0.5372 0.7330
No log 27.125 434 0.5548 0.5044 0.5548 0.7448
No log 27.25 436 0.5711 0.5095 0.5711 0.7557
No log 27.375 438 0.5381 0.5607 0.5381 0.7336
No log 27.5 440 0.5070 0.5656 0.5070 0.7121
No log 27.625 442 0.5207 0.5784 0.5207 0.7216
No log 27.75 444 0.5441 0.6150 0.5441 0.7376
No log 27.875 446 0.5250 0.5597 0.5250 0.7246
No log 28.0 448 0.5032 0.5782 0.5032 0.7094
No log 28.125 450 0.5206 0.6143 0.5206 0.7215
No log 28.25 452 0.5874 0.5378 0.5874 0.7664
No log 28.375 454 0.6114 0.5515 0.6114 0.7819
No log 28.5 456 0.5873 0.5168 0.5873 0.7663
No log 28.625 458 0.5366 0.5621 0.5366 0.7325
No log 28.75 460 0.5155 0.5941 0.5155 0.7180
No log 28.875 462 0.5684 0.5140 0.5685 0.7540
No log 29.0 464 0.5879 0.4920 0.5879 0.7668
No log 29.125 466 0.5631 0.5468 0.5631 0.7504
No log 29.25 468 0.5299 0.6242 0.5299 0.7279
No log 29.375 470 0.5537 0.5136 0.5537 0.7441
No log 29.5 472 0.5624 0.4689 0.5624 0.7499
No log 29.625 474 0.5374 0.5399 0.5374 0.7331
No log 29.75 476 0.5215 0.6228 0.5215 0.7222
No log 29.875 478 0.5353 0.5614 0.5353 0.7317
No log 30.0 480 0.5576 0.5367 0.5576 0.7467
No log 30.125 482 0.5526 0.5367 0.5526 0.7434
No log 30.25 484 0.5351 0.5386 0.5351 0.7315
No log 30.375 486 0.5268 0.5151 0.5268 0.7258
No log 30.5 488 0.5215 0.5404 0.5215 0.7221
No log 30.625 490 0.5202 0.5151 0.5202 0.7212
No log 30.75 492 0.5138 0.5404 0.5138 0.7168
No log 30.875 494 0.5086 0.5404 0.5086 0.7131
No log 31.0 496 0.4932 0.5648 0.4932 0.7023
No log 31.125 498 0.4826 0.5860 0.4826 0.6947
0.2942 31.25 500 0.5007 0.6127 0.5007 0.7076
0.2942 31.375 502 0.5182 0.5855 0.5182 0.7198
0.2942 31.5 504 0.5176 0.6127 0.5176 0.7195
0.2942 31.625 506 0.5045 0.6389 0.5045 0.7103
0.2942 31.75 508 0.4954 0.6073 0.4954 0.7038
0.2942 31.875 510 0.4933 0.6024 0.4933 0.7023
0.2942 32.0 512 0.4973 0.6111 0.4973 0.7052
0.2942 32.125 514 0.5020 0.5897 0.5020 0.7085
0.2942 32.25 516 0.5103 0.6101 0.5103 0.7144

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_usingALLEssays_FineTuningAraBERT_run1_AugV5_k4_task7_organization

Finetuned
(4019)
this model