--- language: - en - code license: apache-2.0 base_model: Qwen/Qwen2.5-Coder-1.5B-Instruct tags: - lora - code - qwen2.5-coder - fingpt - code-correction pipeline_tag: text-generation --- # fingpt-coder-1b5 LoRA adapter for **[Qwen/Qwen2.5-Coder-1.5B-Instruct](https://huggingface.co/Qwen/Qwen2.5-Coder-1.5B-Instruct)** fine-tuned on [m-a-p/Code-Feedback](https://huggingface.co/datasets/m-a-p/Code-Feedback) (66K error→fix pairs, 3 epochs). > **Adapter only** — the base model is loaded from the HF Hub automatically. > Total download: ~84 MB adapter + ~3 GB base model. --- ## LoRA config | Property | Value | |----------|-------| | Base model | `Qwen/Qwen2.5-Coder-1.5B-Instruct` | | Rank (r) | 16 | | Alpha | 32 (scale = 2.0) | | Target modules | `q_proj`, `k_proj`, `v_proj`, `o_proj`, `gate_proj`, `up_proj`, `down_proj` | | Training step | 48500 | | Adapter size | ~84 MB | --- ## Quick start ```bash git clone https://huggingface.co/revana/fingpt-coder-1b5 ``` ```python import torch, sys sys.path.insert(0, "fingpt") # fingpt repo root from infer import load_model, generate model, tokenizer = load_model("adapter_final.pt") reply = generate(model, tokenizer, "Fix this bug:\n\ndef fact(n):\n return n * fact(n)") print(reply) ``` Or use the [live demo](https://huggingface.co/spaces/revana/fingpt). --- ## Training | Property | Value | |----------|-------| | Dataset | [m-a-p/Code-Feedback](https://huggingface.co/datasets/m-a-p/Code-Feedback) | | Samples | ~66K error→fix pairs | | Epochs | 3 | | Batch size | 4 × 4 grad accum = 16 effective | | LR | 3e-4, cosine decay, 3% warmup | | Precision | bfloat16 | | Hardware | A100 80GB | --- ## License Apache 2.0