| license: apache-2.0 | |
| # mistral-7b-alpaca-sft | |
| mistral-7b-alpaca-sft is an SFT fine-tuned version of [unsloth/mistral-7b-bnb-4bit](https://huggingface.co/unsloth/mistral-7b-bnb-4bit) using the [yahma/alpaca-cleaned](https://huggingface.co/datasets/yahma/alpaca-cleaned) dataset. | |
| ## Fine-tuning configuration | |
| ### LoRA | |
| - r: 256 | |
| - LoRA alpha: 128 | |
| - LoRA dropout: 0.0 | |
| ### Training arguments | |
| - Epochs: 1 | |
| - Batch size: 4 | |
| - Gradient accumulation steps: 6 | |
| - Optimizer: adamw_torch_fused | |
| - Max steps: 100 | |
| - Learning rate: 0.0002 | |
| - Weight decay: 0.1 | |
| - Learning rate scheduler type: linear | |
| - Max seq length: 2048 | |
| - 4-bit bnb: True | |
| Trained with [Unsloth](https://github.com/unslothai/unsloth) and Huggingface's TRL library. | |
| [<img src="https://raw.githubusercontent.com/unslothai/unsloth/main/images/unsloth%20made%20with%20love.png" width="200"/>](https://github.com/unslothai/unsloth) |