ORIGINAL_LORA_Model / README.md
jahb57's picture
jahb57/ORIGINAL_LORA_Model
d0eec1f
metadata
license: mit
base_model: jahb57/test_trainer
tags:
  - generated_from_trainer
datasets:
  - yelp_review_full
metrics:
  - accuracy
model-index:
  - name: ORIGINAL_LORA_Model
    results: []

ORIGINAL_LORA_Model

This model is a fine-tuned version of jahb57/test_trainer on the yelp_review_full dataset. It achieves the following results on the evaluation set:

  • Loss: 0.8158
  • Accuracy: 0.6355

Model description

More information needed

Intended uses & limitations

More information needed

Training and evaluation data

More information needed

Training procedure

Training hyperparameters

The following hyperparameters were used during training:

  • learning_rate: 5e-05
  • train_batch_size: 12
  • eval_batch_size: 12
  • seed: 42
  • optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
  • lr_scheduler_type: linear
  • num_epochs: 10

Training results

Training Loss Epoch Step Validation Loss Accuracy
0.9375 1.0 1667 0.8691 0.6155
0.8754 2.0 3334 0.8640 0.619
0.865 3.0 5001 0.8303 0.628
0.8558 4.0 6668 0.8217 0.6295
0.8267 5.0 8335 0.8195 0.6295
0.8042 6.0 10002 0.8388 0.638
0.797 7.0 11669 0.8172 0.6405
0.7887 8.0 13336 0.8146 0.6385
0.7764 9.0 15003 0.8202 0.635
0.7807 10.0 16670 0.8158 0.6355

Framework versions

  • Transformers 4.35.0
  • Pytorch 2.1.0+cu118
  • Datasets 2.14.6
  • Tokenizers 0.14.1