ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k2_task7_organization

This model is a fine-tuned version of aubmindlab/bert-base-arabertv02 on the None dataset. It achieves the following results on the evaluation set:

  • Loss: 0.4239
  • Qwk: 0.5736
  • Mse: 0.4239
  • Rmse: 0.6511

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 2e-05
  • train_batch_size: 8
  • eval_batch_size: 8
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 100

Training results

Training Loss Epoch Step Validation Loss Qwk Mse Rmse
No log 0.2 2 2.6252 -0.1213 2.6252 1.6202
No log 0.4 4 1.0898 0.0989 1.0898 1.0439
No log 0.6 6 0.6940 0.0937 0.6940 0.8331
No log 0.8 8 0.6458 0.2607 0.6458 0.8036
No log 1.0 10 0.6389 0.4026 0.6389 0.7993
No log 1.2 12 0.5658 0.4425 0.5658 0.7522
No log 1.4 14 0.5621 0.4561 0.5621 0.7497
No log 1.6 16 0.8070 0.3620 0.8070 0.8984
No log 1.8 18 0.9283 0.3785 0.9283 0.9635
No log 2.0 20 0.7312 0.4120 0.7312 0.8551
No log 2.2 22 0.4863 0.4729 0.4863 0.6974
No log 2.4 24 0.8927 0.3110 0.8927 0.9448
No log 2.6 26 0.8085 0.4142 0.8085 0.8991
No log 2.8 28 0.4579 0.5422 0.4579 0.6767
No log 3.0 30 0.6131 0.4842 0.6131 0.7830
No log 3.2 32 0.6358 0.5692 0.6358 0.7974
No log 3.4 34 0.4659 0.5528 0.4659 0.6826
No log 3.6 36 0.4211 0.5915 0.4211 0.6489
No log 3.8 38 0.5074 0.5331 0.5074 0.7123
No log 4.0 40 0.6261 0.5358 0.6261 0.7913
No log 4.2 42 0.5683 0.5489 0.5683 0.7539
No log 4.4 44 0.4275 0.6340 0.4275 0.6538
No log 4.6 46 0.4230 0.5914 0.4230 0.6504
No log 4.8 48 0.4247 0.6161 0.4247 0.6517
No log 5.0 50 0.4100 0.6161 0.4100 0.6403
No log 5.2 52 0.4682 0.5875 0.4682 0.6843
No log 5.4 54 0.4595 0.6318 0.4595 0.6779
No log 5.6 56 0.3576 0.6027 0.3576 0.5980
No log 5.8 58 0.3779 0.6706 0.3779 0.6147
No log 6.0 60 0.3729 0.6914 0.3729 0.6106
No log 6.2 62 0.5420 0.5723 0.5420 0.7362
No log 6.4 64 0.5007 0.6169 0.5007 0.7076
No log 6.6 66 0.4015 0.6897 0.4015 0.6336
No log 6.8 68 0.4053 0.6959 0.4053 0.6367
No log 7.0 70 0.4044 0.7080 0.4044 0.6359
No log 7.2 72 0.4026 0.6817 0.4026 0.6345
No log 7.4 74 0.4314 0.6595 0.4314 0.6568
No log 7.6 76 0.5438 0.5922 0.5438 0.7374
No log 7.8 78 0.4474 0.6240 0.4474 0.6689
No log 8.0 80 0.3830 0.7370 0.3830 0.6189
No log 8.2 82 0.3807 0.7431 0.3807 0.6170
No log 8.4 84 0.4474 0.6740 0.4474 0.6689
No log 8.6 86 0.4751 0.6129 0.4751 0.6892
No log 8.8 88 0.3724 0.7209 0.3724 0.6102
No log 9.0 90 0.4265 0.5597 0.4265 0.6531
No log 9.2 92 0.4088 0.5528 0.4088 0.6394
No log 9.4 94 0.3831 0.6552 0.3831 0.6190
No log 9.6 96 0.3753 0.7059 0.3753 0.6126
No log 9.8 98 0.3883 0.7229 0.3883 0.6231
No log 10.0 100 0.3774 0.7260 0.3774 0.6143
No log 10.2 102 0.4028 0.7158 0.4028 0.6346
No log 10.4 104 0.4220 0.7474 0.4220 0.6496
No log 10.6 106 0.5286 0.6589 0.5286 0.7270
No log 10.8 108 0.5777 0.5354 0.5777 0.7601
No log 11.0 110 0.4572 0.7372 0.4572 0.6762
No log 11.2 112 0.3961 0.7267 0.3961 0.6294
No log 11.4 114 0.4868 0.5943 0.4868 0.6977
No log 11.6 116 0.4978 0.5943 0.4978 0.7055
No log 11.8 118 0.4192 0.7057 0.4192 0.6475
No log 12.0 120 0.3892 0.6728 0.3892 0.6238
No log 12.2 122 0.4008 0.6349 0.4008 0.6331
No log 12.4 124 0.5245 0.5363 0.5245 0.7242
No log 12.6 126 0.4675 0.6023 0.4675 0.6837
No log 12.8 128 0.3921 0.6515 0.3921 0.6262
No log 13.0 130 0.4408 0.6796 0.4408 0.6639
No log 13.2 132 0.4702 0.6355 0.4702 0.6857
No log 13.4 134 0.4229 0.6886 0.4229 0.6503
No log 13.6 136 0.4136 0.5611 0.4136 0.6431
No log 13.8 138 0.4286 0.5449 0.4286 0.6547
No log 14.0 140 0.4405 0.4997 0.4405 0.6637
No log 14.2 142 0.4140 0.6271 0.4140 0.6434
No log 14.4 144 0.3978 0.6817 0.3978 0.6307
No log 14.6 146 0.4216 0.6886 0.4216 0.6493
No log 14.8 148 0.4340 0.6824 0.4340 0.6588
No log 15.0 150 0.5119 0.5867 0.5119 0.7155
No log 15.2 152 0.7249 0.4735 0.7249 0.8514
No log 15.4 154 0.6607 0.5169 0.6607 0.8128
No log 15.6 156 0.4630 0.5849 0.4630 0.6805
No log 15.8 158 0.4082 0.6567 0.4082 0.6389
No log 16.0 160 0.4169 0.6370 0.4169 0.6457
No log 16.2 162 0.4211 0.6168 0.4211 0.6489
No log 16.4 164 0.4944 0.5146 0.4944 0.7032
No log 16.6 166 0.5174 0.5146 0.5174 0.7193
No log 16.8 168 0.4263 0.6330 0.4263 0.6529
No log 17.0 170 0.4143 0.6272 0.4143 0.6437
No log 17.2 172 0.4430 0.6150 0.4430 0.6656
No log 17.4 174 0.4105 0.6460 0.4105 0.6407
No log 17.6 176 0.4102 0.6170 0.4102 0.6405
No log 17.8 178 0.4271 0.5046 0.4271 0.6536
No log 18.0 180 0.4576 0.5033 0.4576 0.6765
No log 18.2 182 0.4236 0.5046 0.4236 0.6508
No log 18.4 184 0.4035 0.6477 0.4035 0.6352
No log 18.6 186 0.3974 0.6860 0.3974 0.6304
No log 18.8 188 0.3917 0.6835 0.3917 0.6258
No log 19.0 190 0.4015 0.6867 0.4015 0.6336
No log 19.2 192 0.4367 0.5983 0.4367 0.6608
No log 19.4 194 0.4741 0.5129 0.4741 0.6886
No log 19.6 196 0.4304 0.5283 0.4304 0.6560
No log 19.8 198 0.3922 0.6655 0.3922 0.6262
No log 20.0 200 0.4025 0.6541 0.4025 0.6344
No log 20.2 202 0.3929 0.6650 0.3929 0.6268
No log 20.4 204 0.4028 0.6567 0.4028 0.6347
No log 20.6 206 0.4125 0.6389 0.4125 0.6422
No log 20.8 208 0.4574 0.5599 0.4574 0.6763
No log 21.0 210 0.4155 0.6305 0.4155 0.6446
No log 21.2 212 0.4046 0.6577 0.4046 0.6361
No log 21.4 214 0.4112 0.6807 0.4112 0.6413
No log 21.6 216 0.4087 0.6828 0.4087 0.6393
No log 21.8 218 0.4498 0.5970 0.4498 0.6707
No log 22.0 220 0.4985 0.5528 0.4985 0.7060
No log 22.2 222 0.4639 0.5429 0.4639 0.6811
No log 22.4 224 0.4212 0.5687 0.4212 0.6490
No log 22.6 226 0.4302 0.6639 0.4302 0.6559
No log 22.8 228 0.4288 0.6817 0.4288 0.6548
No log 23.0 230 0.4254 0.64 0.4254 0.6522
No log 23.2 232 0.4251 0.5736 0.4251 0.6520
No log 23.4 234 0.4184 0.6087 0.4184 0.6468
No log 23.6 236 0.4195 0.6301 0.4195 0.6477
No log 23.8 238 0.4226 0.6977 0.4226 0.6501
No log 24.0 240 0.4274 0.6738 0.4274 0.6538
No log 24.2 242 0.4382 0.7059 0.4382 0.6620
No log 24.4 244 0.4803 0.6108 0.4803 0.6931
No log 24.6 246 0.4730 0.5445 0.4730 0.6877
No log 24.8 248 0.4193 0.6245 0.4193 0.6476
No log 25.0 250 0.4431 0.6706 0.4431 0.6657
No log 25.2 252 0.4382 0.6706 0.4382 0.6620
No log 25.4 254 0.4184 0.6446 0.4184 0.6468
No log 25.6 256 0.4227 0.5861 0.4227 0.6502
No log 25.8 258 0.4288 0.5339 0.4288 0.6548
No log 26.0 260 0.4249 0.5098 0.4249 0.6519
No log 26.2 262 0.4219 0.6171 0.4219 0.6495
No log 26.4 264 0.4292 0.6541 0.4292 0.6551
No log 26.6 266 0.4301 0.6185 0.4301 0.6558
No log 26.8 268 0.4343 0.5974 0.4343 0.6590
No log 27.0 270 0.4396 0.5904 0.4396 0.6630
No log 27.2 272 0.4310 0.6222 0.4310 0.6565
No log 27.4 274 0.4213 0.6541 0.4213 0.6491
No log 27.6 276 0.4242 0.6541 0.4242 0.6513
No log 27.8 278 0.4511 0.6956 0.4511 0.6716
No log 28.0 280 0.4279 0.6541 0.4279 0.6542
No log 28.2 282 0.4226 0.6329 0.4226 0.6501
No log 28.4 284 0.4234 0.5841 0.4234 0.6507
No log 28.6 286 0.4131 0.6389 0.4131 0.6428
No log 28.8 288 0.4117 0.6114 0.4117 0.6416
No log 29.0 290 0.3990 0.6053 0.3990 0.6317
No log 29.2 292 0.4068 0.6160 0.4068 0.6378
No log 29.4 294 0.4070 0.6009 0.4070 0.6379
No log 29.6 296 0.4214 0.6671 0.4214 0.6491
No log 29.8 298 0.4339 0.6832 0.4339 0.6587
No log 30.0 300 0.4279 0.6832 0.4279 0.6542
No log 30.2 302 0.4284 0.6569 0.4284 0.6545
No log 30.4 304 0.4434 0.6617 0.4434 0.6659
No log 30.6 306 0.4156 0.6127 0.4156 0.6447
No log 30.8 308 0.4035 0.6146 0.4035 0.6352
No log 31.0 310 0.4202 0.6317 0.4202 0.6483
No log 31.2 312 0.4210 0.6146 0.4210 0.6488
No log 31.4 314 0.4167 0.5974 0.4167 0.6455
No log 31.6 316 0.4313 0.6127 0.4313 0.6568
No log 31.8 318 0.4424 0.5495 0.4424 0.6651
No log 32.0 320 0.4670 0.4841 0.4670 0.6834
No log 32.2 322 0.4411 0.5798 0.4411 0.6642
No log 32.4 324 0.4289 0.6307 0.4289 0.6549
No log 32.6 326 0.4305 0.6076 0.4305 0.6561
No log 32.8 328 0.4288 0.6001 0.4288 0.6548
No log 33.0 330 0.4264 0.6228 0.4264 0.6530
No log 33.2 332 0.4315 0.5707 0.4315 0.6569
No log 33.4 334 0.4328 0.5874 0.4328 0.6579
No log 33.6 336 0.4367 0.5476 0.4367 0.6608
No log 33.8 338 0.4421 0.5677 0.4421 0.6649
No log 34.0 340 0.4411 0.5798 0.4411 0.6641
No log 34.2 342 0.4385 0.6185 0.4385 0.6622
No log 34.4 344 0.4342 0.5826 0.4342 0.6589
No log 34.6 346 0.4334 0.5999 0.4334 0.6583
No log 34.8 348 0.4370 0.5999 0.4370 0.6611
No log 35.0 350 0.4385 0.5569 0.4385 0.6622
No log 35.2 352 0.4234 0.6197 0.4234 0.6507
No log 35.4 354 0.4096 0.6171 0.4096 0.6400
No log 35.6 356 0.4222 0.6310 0.4222 0.6498
No log 35.8 358 0.4374 0.6516 0.4374 0.6613
No log 36.0 360 0.4385 0.6516 0.4385 0.6622
No log 36.2 362 0.4304 0.6516 0.4304 0.6561
No log 36.4 364 0.4292 0.6235 0.4292 0.6551
No log 36.6 366 0.4381 0.5765 0.4381 0.6619
No log 36.8 368 0.4145 0.6673 0.4145 0.6438
No log 37.0 370 0.4021 0.6001 0.4021 0.6341
No log 37.2 372 0.4012 0.6530 0.4012 0.6334
No log 37.4 374 0.3986 0.6745 0.3986 0.6314
No log 37.6 376 0.4033 0.6371 0.4033 0.6350
No log 37.8 378 0.4458 0.6030 0.4458 0.6677
No log 38.0 380 0.4935 0.6081 0.4935 0.7025
No log 38.2 382 0.5083 0.5997 0.5083 0.7130
No log 38.4 384 0.5178 0.5373 0.5178 0.7196
No log 38.6 386 0.4684 0.5348 0.4684 0.6844
No log 38.8 388 0.4424 0.5812 0.4424 0.6651
No log 39.0 390 0.4588 0.6132 0.4588 0.6774
No log 39.2 392 0.4725 0.5923 0.4725 0.6874
No log 39.4 394 0.4620 0.5784 0.4620 0.6797
No log 39.6 396 0.4487 0.5556 0.4487 0.6699
No log 39.8 398 0.4504 0.4802 0.4504 0.6711
No log 40.0 400 0.4647 0.5144 0.4647 0.6817
No log 40.2 402 0.4489 0.5428 0.4489 0.6700
No log 40.4 404 0.4372 0.6551 0.4372 0.6612
No log 40.6 406 0.4439 0.6096 0.4439 0.6663
No log 40.8 408 0.4432 0.6096 0.4432 0.6657
No log 41.0 410 0.4439 0.5538 0.4439 0.6662
No log 41.2 412 0.4513 0.5538 0.4513 0.6718
No log 41.4 414 0.4535 0.5538 0.4535 0.6734
No log 41.6 416 0.4361 0.6053 0.4361 0.6604
No log 41.8 418 0.4410 0.6530 0.4410 0.6641
No log 42.0 420 0.4420 0.6530 0.4420 0.6649
No log 42.2 422 0.4331 0.6530 0.4331 0.6581
No log 42.4 424 0.4295 0.6330 0.4295 0.6554
No log 42.6 426 0.4338 0.5734 0.4338 0.6587
No log 42.8 428 0.4437 0.5092 0.4437 0.6661
No log 43.0 430 0.4514 0.5413 0.4514 0.6719
No log 43.2 432 0.4555 0.5413 0.4555 0.6749
No log 43.4 434 0.4636 0.5614 0.4636 0.6809
No log 43.6 436 0.4691 0.5547 0.4691 0.6849
No log 43.8 438 0.4637 0.5697 0.4637 0.6809
No log 44.0 440 0.4521 0.5406 0.4521 0.6724
No log 44.2 442 0.4446 0.5379 0.4446 0.6668
No log 44.4 444 0.4480 0.6215 0.4480 0.6693
No log 44.6 446 0.4448 0.6215 0.4448 0.6669
No log 44.8 448 0.4335 0.6125 0.4335 0.6584
No log 45.0 450 0.4258 0.6255 0.4258 0.6525
No log 45.2 452 0.4278 0.6492 0.4278 0.6541
No log 45.4 454 0.4324 0.6411 0.4324 0.6576
No log 45.6 456 0.4355 0.5966 0.4355 0.6599
No log 45.8 458 0.4266 0.6411 0.4266 0.6532
No log 46.0 460 0.4165 0.6210 0.4165 0.6454
No log 46.2 462 0.4133 0.6210 0.4133 0.6428
No log 46.4 464 0.4138 0.5985 0.4138 0.6432
No log 46.6 466 0.4108 0.5765 0.4108 0.6409
No log 46.8 468 0.4071 0.5846 0.4071 0.6381
No log 47.0 470 0.4057 0.6448 0.4057 0.6370
No log 47.2 472 0.4103 0.6455 0.4103 0.6405
No log 47.4 474 0.4223 0.6541 0.4223 0.6499
No log 47.6 476 0.4291 0.6541 0.4291 0.6551
No log 47.8 478 0.4293 0.6541 0.4293 0.6552
No log 48.0 480 0.4215 0.6541 0.4215 0.6492
No log 48.2 482 0.4184 0.6001 0.4184 0.6468
No log 48.4 484 0.4245 0.5619 0.4245 0.6515
No log 48.6 486 0.4354 0.5571 0.4354 0.6598
No log 48.8 488 0.4386 0.5826 0.4386 0.6623
No log 49.0 490 0.4331 0.5826 0.4331 0.6581
No log 49.2 492 0.4243 0.5687 0.4243 0.6514
No log 49.4 494 0.4227 0.5722 0.4227 0.6501
No log 49.6 496 0.4235 0.5722 0.4235 0.6507
No log 49.8 498 0.4213 0.6092 0.4213 0.6491
0.1947 50.0 500 0.4234 0.6092 0.4234 0.6507
0.1947 50.2 502 0.4248 0.5860 0.4248 0.6518
0.1947 50.4 504 0.4248 0.5640 0.4248 0.6518
0.1947 50.6 506 0.4263 0.6108 0.4263 0.6529
0.1947 50.8 508 0.4278 0.6034 0.4278 0.6540
0.1947 51.0 510 0.4376 0.5509 0.4376 0.6615
0.1947 51.2 512 0.4448 0.5467 0.4448 0.6669
0.1947 51.4 514 0.4316 0.5663 0.4316 0.6569
0.1947 51.6 516 0.4239 0.5736 0.4239 0.6511

Framework versions

  • Transformers 4.44.2
  • Pytorch 2.4.0+cu118
  • Datasets 2.21.0
  • Tokenizers 0.19.1
Downloads last month
-
Safetensors
Model size
0.1B params
Tensor type
F32
·
Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for MayBashendy/ArabicNewSplits7_OSS_usingWellWrittenEssays_FineTuningAraBERT_run1_AugV5_k2_task7_organization

Finetuned
(4019)
this model