|
|
--- |
|
|
library_name: transformers |
|
|
pipeline_tag: text-generation |
|
|
tags: |
|
|
- synthetic-data |
|
|
- dpo |
|
|
- gpqa |
|
|
- reasoning |
|
|
- alignment |
|
|
- quantum |
|
|
- neuroscience |
|
|
- gloss-free |
|
|
- data-efficient |
|
|
base_model: Qwen/Qwen2.5-7B-Instruct |
|
|
license: other |
|
|
language: |
|
|
- en |
|
|
metrics: |
|
|
- accuracy |
|
|
datasets: |
|
|
- TrueRunAI/TrueRun-Groove-v2.1-DPO |
|
|
--- |
|
|
|
|
|
# TrueRun-Groove-v2.1-7B |
|
|
|
|
|
Qwen2.5-7B-Instruct fine-tuned on ~1,200 high-rigor synthetic DPO pairs (Groove v2.1). |
|
|
|
|
|
Balanced quantum mechanics, neuroscience/BCI, alignment/game theory. Structural escalation for indefinite depth—no gloss decay. |
|
|
|
|
|
## Key Results (GPQA Diamond, 3 Seeds Mean) |
|
|
| Benchmark | Questions | Baseline % | Groove Mean % | Delta | Notes | |
|
|
|--------------------|-----------|------------|---------------|-----------|-------| |
|
|
| Full Diamond | 198 | 33.33% | 36.53% | +3.20% | Low variance (±0.58%) | |
|
|
| Quantum Subset | 39 | 35.90% | 51.92% | +16.02% | Leading public targeted lift for 7B | |
|
|
| Biology Subset | 19 | 36.84% | 52.63% | +15.79% | Strong transfer | |
|
|
| Physics Subset | 86 | 51.16% | 42.25% | -8.91% | Targeted regression—next iter fix | |
|
|
|
|
|
Leading data efficiency & domain-specific gains among public 7B fine-tunes. |
|
|
|
|
|
## License |
|
|
Other (non-exclusive commercial/research use—dataset for sale on OpenDataBay; model weights public for testing/reproduction). |
|
|
|
|
|
## Usage |
|
|
```python |
|
|
from transformers import pipeline |
|
|
|
|
|
pipe = pipeline("text-generation", model="TrueRunAI/TrueRun-Groove-v2.1-7B") |
|
|
pipe("Explain quantum entanglement simply but without losing rigor:", max_new_tokens=256) |