|
|
--- |
|
|
base_model: sentence-transformers/all-MiniLM-L6-v2 |
|
|
library_name: peft |
|
|
license: mit |
|
|
tags: |
|
|
- lora |
|
|
- peft |
|
|
- scientific |
|
|
- research |
|
|
- academic |
|
|
- domain-adaptation |
|
|
- sentence-embeddings |
|
|
language: |
|
|
- en |
|
|
--- |
|
|
|
|
|
# Scientific LoRA Adapter for DomainEmbedder-v2.6 |
|
|
|
|
|
Domain-specific LoRA adapter for scientific/research text embeddings. |
|
|
|
|
|
## Model Details |
|
|
|
|
|
| Property | Value | |
|
|
|----------|-------| |
|
|
| **Base Model** | sentence-transformers/all-MiniLM-L6-v2 | |
|
|
| **Parent System** | DomainEmbedder-v2.6 | |
|
|
| **Domain** | Scientific / Research | |
|
|
| **LoRA Rank** | 16 | |
|
|
| **LoRA Alpha** | 32 | |
|
|
| **Target Modules** | query, value | |
|
|
| **Trainable Params** | 147,456 (0.645%) | |
|
|
|
|
|
## Training Data |
|
|
|
|
|
Trained on 40,000 scientific text pairs from: |
|
|
- arXiv (document-level) |
|
|
- arXiv (section-level) |
|
|
- PubMed Artificial |
|
|
- Scientific Papers |
|
|
|
|
|
**Note**: 87.3% real data + 12.7% augmented data (scientific domain had fewer available pairs) |
|
|
|
|
|
## Training Configuration |
|
|
|
|
|
| Parameter | Value | |
|
|
|-----------|-------| |
|
|
| Epochs | 3 | |
|
|
| Batch Size | 32 | |
|
|
| Learning Rate | 2e-4 | |
|
|
| Loss | Contrastive (InfoNCE) | |
|
|
| Best Val Loss | 0.0016 | |
|
|
|
|
|
## Usage |
|
|
|
|
|
This adapter is part of the DomainEmbedder-v2.6 system. It is selected automatically by the RL policy when scientific content is detected. |
|
|
|
|
|
```python |
|
|
from peft import PeftModel |
|
|
from transformers import AutoModel |
|
|
|
|
|
# Load base encoder |
|
|
base_encoder = AutoModel.from_pretrained('sentence-transformers/all-MiniLM-L6-v2') |
|
|
|
|
|
# Apply scientific LoRA |
|
|
scientific_model = PeftModel.from_pretrained(base_encoder, 'path/to/scientific_lora') |
|
|
``` |
|
|
|
|
|
## Author |
|
|
|
|
|
**Zain Asad** |
|
|
|
|
|
## License |
|
|
|
|
|
MIT License |
|
|
|
|
|
## Framework Versions |
|
|
|
|
|
- PEFT 0.18.1 |
|
|
- Transformers 4.x |
|
|
- PyTorch 2.x |
|
|
|