--- base_model: Qwen/Qwen2.5-Coder-14B library_name: peft pipeline_tag: text-generation license: apache-2.0 tags: - code - coding-assistant - lora - refactor - bug-fix - optimization - async - concurrency - security - logging - networking - transformers --- # EdgePulse Coder 14B (LoRA) **EdgePulse Coder 14B** is a production-grade coding assistant fine-tuned using LoRA on top of **Qwen2.5-Coder-14B**. It is designed to handle real-world software engineering workflows with high reliability and correctness. --- ## Model Details ### Model Description EdgePulse Coder 14B focuses on **practical developer tasks**, trained on a large, strictly validated dataset covering: - Bug fixing - Code explanation - Refactoring - Optimization - Async & concurrency correction - Logging & observability - Security & defensive coding - Networking & I/O handling - Multi-file context reasoning - Test generation and impact analysis The model is optimized for **IDE usage**, **CLI workflows**, and **Cursor-like streaming environments**. --- - **Developed by:** EdgePulseAI - **Shared by:** EdgePulseAI - **Model type:** Large Language Model (Code-focused) - **Language(s):** Python, JavaScript, TypeScript, Bash (primary), general programming concepts - **License:** Apache-2.0 - **Finetuned from:** Qwen/Qwen2.5-Coder-14B --- ## Model Sources - **Base Model:** https://huggingface.co/Qwen/Qwen2.5-Coder-14B - **Website:** https://EdgePulseAi.com --- ## Uses ### Direct Use EdgePulse Coder 14B can be used directly for: - Code explanation - Bug fixing - Refactoring existing code - Generating tests - Improving logging and error handling - Fixing async / concurrency bugs - Secure coding suggestions - Network & I/O robustness ### Downstream Use - IDE assistants (VS Code / Cursor-style tools) - CI/CD automation - Code review bots - Developer copilots - Internal engineering tools ### Out-of-Scope Use - Medical or legal advice - Autonomous system control - High-risk decision making without human review --- ## Bias, Risks, and Limitations - The model may occasionally produce syntactically correct but logically incorrect code. - Security-sensitive code should always be reviewed by humans. - Performance depends on correct prompt framing and context size. ### Recommendations - Use human review for production deployments. - Combine with static analysis and testing tools. - Prefer structured prompts for multi-file tasks. --- ## How to Get Started with the Model ```python from transformers import AutoTokenizer, AutoModelForCausalLM from peft import PeftModel base_model = "Qwen/Qwen2.5-Coder-14B" adapter_model = "edgepulse-ai/EdgePulse-Coder-14B-LoRA" tokenizer = AutoTokenizer.from_pretrained(base_model) model = AutoModelForCausalLM.from_pretrained( base_model, device_map="auto", torch_dtype="auto" ) model = PeftModel.from_pretrained(model, adapter_model) model.eval() prompt = "Fix this bug:\n\ndef add(a,b): return a-b" inputs = tokenizer(prompt, return_tensors="pt").to(model.device) output = model.generate(**inputs, max_new_tokens=128) print(tokenizer.decode(output[0], skip_special_tokens=True))