External Communication Guide: FDRA Long-Context Results
Date: 2026-01-22
Purpose: How to frame these results accurately without overreach
The Core Principle
Claim only what the evidence supports. No more.
This guide separates what you can say confidently from what would be overclaiming.
β What You CAN Say (High Confidence)
Technically Accurate Claims
"We identified and fixed Ο collapse during FDRA training"
- Evidence: Half-life incentives + hard constraint maintain Ο distribution
- Logged metrics show stable Ο throughout training
"Routing into slow channels improves identity retention"
- Evidence: Ο-weighted routing outperforms uniform routing on retention probes
- Measured at multiple interference lengths
"Extended Ο range handles longer Gaussian interference"
- Evidence: Failure point shifts from KβΟ_max to Kβ4ΓΟ_max
- Matches theoretical prediction
"Multi-head encoding improves structured interference resistance"
- Evidence: ISA shifts failure from K=512 to K=2048
- Invariant core alignment measured
"Language-level probes show commitment adherence improvement"
- Evidence: 0% β 5% β 40% pass rate across conditions
- Early commitment honored in final output
Safe Summary Statements
"We've shown that FDRA-style architectures can stably preserve long-timescale internal state under realistic training conditions."
"The architectural mechanisms for identity preservation are now validated."
"Remaining limitations appear to arise from task-level supervision, not memory collapse."
β οΈ What You SHOULD NOT Say
Overclaims to Avoid
β "We solved long-context reasoning"
- Reality: We validated memory preservation, not full reasoning
β "FDRA now handles full document understanding"
- Reality: Probes test identity/commitment, not semantic comprehension
β "This works at GPT-4 scale"
- Reality: Validated at toy scale only (32 oscillators, 16 dims)
β "The long-context problem is solved"
- Reality: The architectural question is answered; task-level challenges remain
β "ISA outperforms transformers on long-context"
- Reality: No direct comparison with attention-based architectures
Why These Matter
Overclaiming damages credibility and invites scrutiny you can't withstand. The results are good enough to stand on their own merit without inflation.
π Recommended Phrasing by Context
For Technical Papers
"We demonstrate that half-life regularization and Ο-weighted routing enable FDRA oscillator banks to preserve identity-level information across contexts exceeding 4096 tokens. Multi-head encoding further extends resistance to structured interference. Language-level probes confirm that preserved state governs downstream behavior."
For Internal Discussion
"We resolved the architectural question Melanie raised. Ο collapse can be prevented, and the preserved state is functionally useful. The remaining work is task design and scaling."
For External Collaborators
"We've completed a systematic study of long-context preservation in FDRA architectures. The results validate that the memory substrate works as theorized when trained with appropriate incentives. We're now moving to task-level validation."
For Public Communication
"New results on long-context memory in recurrent architectures. We identified why models forget over long contexts and developed mechanisms to prevent it. Early-commitment probes show 40% improvement in commitment adherence."
π― Key Differentiators
What makes these results legitimate (emphasize these):
- Clean experimental design β Control vs treatment, same seeds, same data
- Mechanistic understanding β Each fix addresses a specific cause
- No oracle cheating β No privileged readout, no rotation inversion
- Language-level validation β Not just synthetic retention metrics
- Internal consistency β Ο distribution, routing, and probes all align
π Numbers You Can Quote
| Metric | Baseline | Routing+HL | ISA | Context |
|---|---|---|---|---|
| Structured interference failure | K=512 | K=512 | K=2048 | 3Γ improvement |
| Gaussian interference failure | K=4096 | K=4096 | K=8192 | 2Γ improvement |
| Language commitment pass rate | 0% | 5% | 40% | 8Γ improvement |
| Ο distribution stability | Collapses | Stable | Stable | β |
β Questions You Should Be Ready to Answer
"How does this compare to attention-based approaches?"
- "We haven't done direct comparison. This validates the FDRA substrate specifically."
"Does this work at real scale?"
- "Validated at toy scale. Scale-up is next."
"Is long-context 'solved'?"
- "The architectural mechanisms are validated. Task-level challenges remain."
"What's the remaining bottleneck?"
- "Credit assignment and readout learning, not memory decay."
"Can I use this in production?"
- "Integration code is available. Validation at your scale needed."
Final Framing Advice
Do This
- Be specific about what was measured
- Acknowledge limitations upfront
- Use "validated" not "solved"
- Distinguish architecture from full system
Don't Do This
- Imply broader claims than evidence supports
- Hide scale limitations
- Conflate retention metrics with reasoning
- Overstate language-level results
One-Paragraph Public Statement (Template)
We present a systematic study of long-context preservation in FDRA recurrent architectures. We identified four mechanisms causing long-context failure and developed targeted fixes: half-life regularization prevents Ο collapse, Ο-weighted routing ensures slow channels are used, extended Ο range handles Gaussian interference, and multi-head encoding (ISA) resists structured overwrite. Language-level probes confirm that early-context commitments are honored in downstream outputs, with 40% pass rate vs 0% baseline. The architectural substrate is now validated; remaining work focuses on task-level supervision and scaling. Code and results at [HuggingFace link].
Remember: The results are good. You don't need to oversell them.