File size: 1,664 Bytes
485908e
 
 
f93132a
485908e
 
f93132a
 
 
 
 
 
 
 
485908e
 
f93132a
485908e
f93132a
485908e
 
 
f93132a
 
 
 
 
 
 
 
 
485908e
f93132a
485908e
f93132a
 
 
 
 
485908e
f93132a
485908e
f93132a
485908e
f93132a
 
 
 
 
 
 
485908e
f93132a
485908e
f93132a
485908e
f93132a
 
 
485908e
f93132a
 
485908e
f93132a
 
 
485908e
f93132a
485908e
f93132a
485908e
f93132a
485908e
f93132a
485908e
f93132a
485908e
f93132a
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
---
base_model: sentence-transformers/all-MiniLM-L6-v2
library_name: peft
license: mit
tags:
- lora
- peft
- scientific
- research
- academic
- domain-adaptation
- sentence-embeddings
language:
- en
---

# Scientific LoRA Adapter for DomainEmbedder-v2.6

Domain-specific LoRA adapter for scientific/research text embeddings.

## Model Details

| Property | Value |
|----------|-------|
| **Base Model** | sentence-transformers/all-MiniLM-L6-v2 |
| **Parent System** | DomainEmbedder-v2.6 |
| **Domain** | Scientific / Research |
| **LoRA Rank** | 16 |
| **LoRA Alpha** | 32 |
| **Target Modules** | query, value |
| **Trainable Params** | 147,456 (0.645%) |

## Training Data

Trained on 40,000 scientific text pairs from:
- arXiv (document-level)
- arXiv (section-level)
- PubMed Artificial
- Scientific Papers

**Note**: 87.3% real data + 12.7% augmented data (scientific domain had fewer available pairs)

## Training Configuration

| Parameter | Value |
|-----------|-------|
| Epochs | 3 |
| Batch Size | 32 |
| Learning Rate | 2e-4 |
| Loss | Contrastive (InfoNCE) |
| Best Val Loss | 0.0016 |

## Usage

This adapter is part of the DomainEmbedder-v2.6 system. It is selected automatically by the RL policy when scientific content is detected.

```python
from peft import PeftModel
from transformers import AutoModel

# Load base encoder
base_encoder = AutoModel.from_pretrained('sentence-transformers/all-MiniLM-L6-v2')

# Apply scientific LoRA
scientific_model = PeftModel.from_pretrained(base_encoder, 'path/to/scientific_lora')
```

## Author

**Zain Asad**

## License

MIT License

## Framework Versions

- PEFT 0.18.1
- Transformers 4.x
- PyTorch 2.x