| --- |
| license: apache-2.0 |
| base_model: Qwen/Qwen2.5-Coder-7B-Instruct |
| tags: |
| - qlora |
| - peft |
| - fine-tune |
| - code-optimization |
| - qwen |
| - code-generation |
| - llm |
| --- |
| |
| # Code Optimization Fine-tuned Qwen2.5-Coder-7B-Instruct (LoRA Adapter) |
|
|
| This repository contains a fine-tuned LoRA adapter for the `Qwen/Qwen2.5-Coder-7B-Instruct` model, specialized for Python code optimization. The model was fine-tuned using QLoRA on the `SeifElden2342532/Code-Optimization` dataset. |
|
|
| ## Model Description |
|
|
| This model is a LoRA adapter of the `Qwen2.5-Coder-7B-Instruct` base model. It has been fine-tuned to act as an expert Python code optimizer, capable of taking user-provided Python code and optimizing it for performance, readability, or conciseness, along with providing explanations and complexity comparisons. |
|
|
| ## Training Details |
|
|
| * **Base Model**: `Qwen/Qwen2.5-Coder-7B-Instruct` |
| * **Fine-tuning Framework**: QLoRA (Parameter-Efficient Fine-Tuning) |
| * **Dataset**: `SeifElden2342532/Code-Optimization` |
| * **Training Parameters:** |
| * `num_train_epochs`: 3 |
| * `lora_r`: 64 |
| * `lora_alpha`: 128 |
| * `max_seq_length`: 2048 |
| * `per_device_train_batch_size`: 4 |
| * `gradient_accumulation_steps`: 4 |
| * `learning_rate`: 2e-4 |
| * `bf16`: True |
| * `gradient_checkpointing`: True |
|
|
| ## Github Repo |
| https://github.com/SeifEldenOsama/Code-Optimizer/tree/master |
|
|
| ## How to Use |
|
|
| To use this LoRA adapter, you need to load the base model and then apply the adapter. Here's a complete example to test the model: |
|
|
| ```python |
| from transformers import AutoTokenizer, AutoModelForCausalLM |
| from peft import PeftModel |
| import torch |
| |
| # 1. Configuration |
| base_model_id = "Qwen/Qwen2.5-Coder-7B-Instruct" |
| adapter_repo_id = "SeifElden2342532/Code-Optimizer" |
| |
| # 2. Load Tokenizer and Base Model |
| tokenizer = AutoTokenizer.from_pretrained(base_model_id) |
| model = AutoModelForCausalLM.from_pretrained( |
| base_model_id, |
| torch_dtype=torch.bfloat16, |
| device_map="auto" |
| ) |
| |
| # 3. Load and Merge the LoRA Adapter |
| model = PeftModel.from_pretrained(model, adapter_repo_id) |
| model = model.merge_and_unload() # Merging for faster inference |
| |
| # 4. Prepare the Input |
| messages = [ |
| { |
| "role": "system", |
| "content": "You are an expert Python code optimizer. Your goal is to take user-provided Python code and optimize it for performance, readability, or conciseness, based on the user's specified category. Provide the optimized code, a brief explanation of the changes, and a complexity comparison table (e.g., time and space complexity before and after optimization)." |
| }, |
| { |
| "role": "user", |
| "content": "Original Code:\n```python\ndef factorial(n):\n if n == 0:\n return 1\n else:\n return n * factorial(n-1)\n```\nCategory: Performance" |
| } |
| ] |
| |
| # 5. Generate Optimization |
| text = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True) |
| model_inputs = tokenizer([text], return_tensors="pt").to(model.device) |
| |
| generated_ids = model.generate(**model_inputs, max_new_tokens=1024) |
| output = tokenizer.batch_decode(generated_ids, skip_special_tokens=True)[0] |
| |
| # 6. Print Result |
| print(output) |