--- title: AI Coding Assistant emoji: ⚡ colorFrom: blue colorTo: purple sdk: docker app_port: 7860 pinned: false license: mit --- # AI Coding Assistant Production-grade RAG-based coding assistant with LangChain, FAISS, and LoRA-tuned LLMs. ## Features - Semantic Code Search with FAISS vector store - LangChain RAG pipeline for context-aware responses - DeepSeek Coder LLM with optional LoRA fine-tuning - 8-bit quantization for efficient inference (GPU only) - Optional Qdrant Cloud integration ## Usage 1. Enter your code repository path (or use sample data) 2. Click "Index Repository" to process your codebase 3. Ask questions or request code fixes 4. Optionally enable LoRA-tuned model (requires GPU) ## Configuration For Qdrant Cloud integration, add secrets in Space settings: - `QDRANT_URL`: Your cluster URL - `QDRANT_API_KEY`: Your API key ## Performance - CPU inference: ~10-30s per response (free tier) - GPU inference: ~2-5s per response (upgrade required) - First load: ~2-3 minutes (model download) ## Local Development ```bash git clone https://github.com/Kash6/localCopilot cd localCopilot # Windows with GPU setup_conda_gpu.bat run_conda.bat # Or use pip pip install -r requirements.txt streamlit run app.py ``` ## Architecture - **Vector Store**: FAISS (default) or Qdrant Cloud (optional) - **Embeddings**: all-MiniLM-L6-v2 - **Reranker**: ms-marco-MiniLM-L-6-v2 - **LLM**: DeepSeek Coder 1.3B - **Framework**: LangChain + Streamlit