--- base_model: sentence-transformers/all-MiniLM-L6-v2 library_name: peft license: mit tags: - lora - peft - code - programming - software - domain-adaptation - sentence-embeddings language: - en --- # Code LoRA Adapter for DomainEmbedder-v2.6 Domain-specific LoRA adapter for code/programming text embeddings. ## Model Details | Property | Value | |----------|-------| | **Base Model** | sentence-transformers/all-MiniLM-L6-v2 | | **Parent System** | DomainEmbedder-v2.6 | | **Domain** | Code / Programming | | **LoRA Rank** | 16 | | **LoRA Alpha** | 32 | | **Target Modules** | query, value | | **Trainable Params** | 147,456 (0.645%) | ## Training Data Trained on 40,000 code-related text pairs from: - Code Alpaca - MBPP (Mostly Basic Python Problems) - Code Contests - Python Instructions ## Training Configuration | Parameter | Value | |-----------|-------| | Epochs | 3 | | Batch Size | 32 | | Learning Rate | 2e-4 | | Loss | Contrastive (InfoNCE) | | Best Val Loss | 0.0039 | ## Usage This adapter is part of the DomainEmbedder-v2.6 system. It is selected automatically by the RL policy when code-related content is detected. ```python from peft import PeftModel from transformers import AutoModel # Load base encoder base_encoder = AutoModel.from_pretrained('sentence-transformers/all-MiniLM-L6-v2') # Apply code LoRA code_model = PeftModel.from_pretrained(base_encoder, 'path/to/code_lora') ``` ## Author **Zain Asad** ## License MIT License ## Framework Versions - PEFT 0.18.1 - Transformers 4.x - PyTorch 2.x