|
|
--- |
|
|
license: apache-2.0 |
|
|
base_model: Qwen/Qwen2.5-Coder-14B |
|
|
tags: |
|
|
- code |
|
|
- qwen2.5 |
|
|
- lora-merged |
|
|
- fine-tuned |
|
|
library_name: transformers |
|
|
--- |
|
|
|
|
|
# PoPilot - Fine-tuned Qwen2.5-Coder-14B |
|
|
|
|
|
This model is a fine-tuned version of [Qwen/Qwen2.5-Coder-14B](https://huggingface.co/Qwen/Qwen2.5-Coder-14B) with LoRA adapters merged. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
- **Base Model**: Qwen/Qwen2.5-Coder-14B |
|
|
- **Fine-tuning Method**: LoRA (Low-Rank Adaptation) |
|
|
- **Training**: Supervised Fine-Tuning (SFT) |
|
|
- **Merged**: Full model weights (LoRA merged with base) |
|
|
|
|
|
## Usage |
|
|
|
|
|
```python |
|
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
|
|
model = AutoModelForCausalLM.from_pretrained( |
|
|
"Justin6657/PoPilot", |
|
|
torch_dtype="auto", |
|
|
device_map="auto", |
|
|
trust_remote_code=True |
|
|
) |
|
|
|
|
|
tokenizer = AutoTokenizer.from_pretrained( |
|
|
"Justin6657/PoPilot", |
|
|
trust_remote_code=True |
|
|
) |
|
|
|
|
|
# Example usage |
|
|
prompt = "Write a Python function to calculate fibonacci numbers:" |
|
|
inputs = tokenizer(prompt, return_tensors="pt") |
|
|
outputs = model.generate(**inputs, max_length=200, temperature=0.7) |
|
|
response = tokenizer.decode(outputs[0], skip_special_tokens=True) |
|
|
print(response) |
|
|
``` |
|
|
|
|
|
## Training Details |
|
|
|
|
|
This model was fine-tuned using LoRA adapters and then merged back into the full model weights. |
|
|
Original LoRA checkpoint path: `/net/projects/CLS/DSI_clinic/justin/checkpoint/augmented_train_Qwen2.5-Coder-14B_full-model_repair-synth_repair-simple-phase4` |
|
|
|