tinyllama-lora-eval / README.md
hassan7272's picture
Upload README.md with huggingface_hub
53f7326 verified
---
license: apache-2.0
base_model: TinyLlama/TinyLlama-1.1B-Chat-v1.0
tags:
- lora
- peft
- fine-tuning
- evaluation
- transformers
---
# TinyLlama LoRA Fine-Tuning (Evaluation Project)
This repository contains **LoRA adapters** fine-tuned on top of
`TinyLlama/TinyLlama-1.1B-Chat-v1.0` using Hugging Face PEFT.
## πŸ”§ Training Details
- **Method**: LoRA (Parameter-Efficient Fine-Tuning)
- **Trainable parameters**: <1% of base model
- **Trainer**: TRL SFTTrainer
- **Hardware**: Google Colab (single GPU)
- **Epochs**: 2
## πŸ“Š Evaluation
The model was evaluated against the base model using identical prompts.
Results showed:
- Noticeable changes in response style and length
- Improved instruction adherence in some cases
- Sensitivity to dataset quality and size
This project focuses on **understanding LoRA behavior and evaluation**
rather than maximizing benchmark scores.
## ⚠️ Limitations
- Small base model (1.1B parameters)
- Limited fine-tuning dataset
- Some responses may be inaccurate or off-topic
## πŸš€ Usage
Example usage (see Hugging Face page for full snippet).
## πŸ“Œ Disclaimer
This repository is for **educational and evaluation purposes**.