Code-Optimizer / README.md
SeifElden2342532's picture
Update README.md
7a0fbc3 verified
---
license: apache-2.0
base_model: Qwen/Qwen2.5-Coder-7B-Instruct
tags:
- qlora
- peft
- fine-tune
- code-optimization
- qwen
- code-generation
- llm
---
# Code Optimization Fine-tuned Qwen2.5-Coder-7B-Instruct (LoRA Adapter)
This repository contains a fine-tuned LoRA adapter for the `Qwen/Qwen2.5-Coder-7B-Instruct` model, specialized for Python code optimization. The model was fine-tuned using QLoRA on the `SeifElden2342532/Code-Optimization` dataset.
## Model Description
This model is a LoRA adapter of the `Qwen2.5-Coder-7B-Instruct` base model. It has been fine-tuned to act as an expert Python code optimizer, capable of taking user-provided Python code and optimizing it for performance, readability, or conciseness, along with providing explanations and complexity comparisons.
## Training Details
* **Base Model**: `Qwen/Qwen2.5-Coder-7B-Instruct`
* **Fine-tuning Framework**: QLoRA (Parameter-Efficient Fine-Tuning)
* **Dataset**: `SeifElden2342532/Code-Optimization`
* **Training Parameters:**
* `num_train_epochs`: 3
* `lora_r`: 64
* `lora_alpha`: 128
* `max_seq_length`: 2048
* `per_device_train_batch_size`: 4
* `gradient_accumulation_steps`: 4
* `learning_rate`: 2e-4
* `bf16`: True
* `gradient_checkpointing`: True
## Github Repo
https://github.com/SeifEldenOsama/Code-Optimizer/tree/master
## How to Use
To use this LoRA adapter, you need to load the base model and then apply the adapter. Here's a complete example to test the model:
```python
from transformers import AutoTokenizer, AutoModelForCausalLM
from peft import PeftModel
import torch
# 1. Configuration
base_model_id = "Qwen/Qwen2.5-Coder-7B-Instruct"
adapter_repo_id = "SeifElden2342532/Code-Optimizer"
# 2. Load Tokenizer and Base Model
tokenizer = AutoTokenizer.from_pretrained(base_model_id)
model = AutoModelForCausalLM.from_pretrained(
base_model_id,
torch_dtype=torch.bfloat16,
device_map="auto"
)
# 3. Load and Merge the LoRA Adapter
model = PeftModel.from_pretrained(model, adapter_repo_id)
model = model.merge_and_unload() # Merging for faster inference
# 4. Prepare the Input
messages = [
{
"role": "system",
"content": "You are an expert Python code optimizer. Your goal is to take user-provided Python code and optimize it for performance, readability, or conciseness, based on the user's specified category. Provide the optimized code, a brief explanation of the changes, and a complexity comparison table (e.g., time and space complexity before and after optimization)."
},
{
"role": "user",
"content": "Original Code:\n```python\ndef factorial(n):\n if n == 0:\n return 1\n else:\n return n * factorial(n-1)\n```\nCategory: Performance"
}
]
# 5. Generate Optimization
text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
model_inputs = tokenizer([text], return_tensors="pt").to(model.device)
generated_ids = model.generate(**model_inputs, max_new_tokens=1024)
output = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0]
# 6. Print Result
print(output)