|
|
--- |
|
|
language: en |
|
|
license: apache-2.0 |
|
|
tags: |
|
|
- llama2 |
|
|
- lora |
|
|
- code |
|
|
- adapter |
|
|
datasets: |
|
|
- iamtarun/python_code_instructions_18k_alpaca |
|
|
--- |
|
|
|
|
|
# LLaMA-2-7B CODE LoRA Adapter |
|
|
|
|
|
This is a LoRA adapter for LLaMA-2-7B fine-tuned on code domain data. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
- **Base Model**: meta-llama/Llama-2-7b-hf |
|
|
- **Adapter Type**: LoRA (Low-Rank Adaptation) |
|
|
- **Domain**: Code |
|
|
- **Training Data**: Python code instructions |
|
|
- **Training Examples**: 2000 (1600 train, 200 val, 200 test) |
|
|
- **Epochs**: 2 |
|
|
|
|
|
## LoRA Configuration |
|
|
|
|
|
- Rank (r): 16 |
|
|
- Alpha: 32 |
|
|
- Dropout: 0.05 |
|
|
- Target Modules: q_proj, v_proj, k_proj, o_proj |
|
|
|
|
|
## Performance Metrics (100 test examples) |
|
|
|
|
|
- Loss: 0.573 |
|
|
- Perplexity: 1.773 |
|
|
- BLEU: 32.76 |
|
|
|
|
|
## Usage |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
from peft import PeftModel |
|
|
|
|
|
# Load base model |
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
"meta-llama/Llama-2-7b-hf", |
|
|
load_in_8bit=True, |
|
|
device_map="auto" |
|
|
) |
|
|
tokenizer = AutoTokenizer.from_pretrained("meta-llama/Llama-2-7b-hf") |
|
|
|
|
|
# Load adapter |
|
|
model = PeftModel.from_pretrained(model, "Thamirawaran/llama2-7b-code-lora") |
|
|
|
|
|
# Generate |
|
|
prompt = 'Write a Python function to sort a list' |
|
|
inputs = tokenizer(prompt, return_tensors="pt").to("cuda") |
|
|
outputs = model.generate(**inputs, max_length=256, temperature=0.7) |
|
|
print(tokenizer.decode(outputs[0], skip_special_tokens=True)) |
|
|
``` |
|
|
|
|
|
## Training Details |
|
|
|
|
|
- Trained with FYP_MDLE library |
|
|
- 8-bit quantization during training |
|
|
- Gradient accumulation: 16 steps |
|
|
- Learning rate: 2e-4 |
|
|
- Warmup steps: 20 |
|
|
|
|
|
## Citation |
|
|
|
|
|
```bibtex |
|
|
@misc{llama2-code-lora, |
|
|
author = {Team RAISE}, |
|
|
title = {LLaMA-2-7B Code LoRA Adapter}, |
|
|
year = {2026}, |
|
|
publisher = {HuggingFace}, |
|
|
howpublished = {\url{https://huggingface.co/Thamirawaran/llama2-7b-code-lora}} |
|
|
} |
|
|
``` |
|
|
|
|
|
## License |
|
|
|
|
|
Apache 2.0 |
|
|
|