Abstract
ATANT presents an open framework for evaluating AI system continuity through a 10-checkpoint methodology using a 250-story corpus across 6 life domains, achieving 100% accuracy in cumulative testing.
We present ATANT (Automated Test for Acceptance of Narrative Truth), an open evaluation framework for measuring continuity in AI systems: the ability to persist, update, disambiguate, and reconstruct meaningful context across time. While the AI industry has produced memory components (RAG pipelines, vector databases, long context windows, profile layers), no published framework formally defines or measures whether these components produce genuine continuity. We define continuity as a system property with 7 required properties, introduce a 10-checkpoint evaluation methodology that operates without an LLM in the evaluation loop, and present a narrative test corpus of 250 stories comprising 1,835 verification questions across 6 life domains. We evaluate a reference implementation across 5 test suite iterations, progressing from 58% (legacy architecture) to 100% in isolated mode (250 stories) and 100% in 50-story cumulative mode, with 96% at 250-story cumulative scale. The cumulative result is the primary measure: when 250 distinct life narratives coexist in the same database, the system must retrieve the correct fact for the correct context without cross-contamination. ATANT is system-agnostic, model-independent, and designed as a sequenced methodology for building and validating continuity systems. The framework specification, example stories, and evaluation protocol are available at https://github.com/Kenotic-Labs/ATANT. The full 250-story corpus will be released incrementally.
Community
ATANT defines continuity as a system property with 7 required properties and introduces the first LLM-free evaluation methodology for AI continuity systems. Includes a 250-story narrative corpus and 10-checkpoint evaluation protocol. Reference implementation reaches 96% at 250-story cumulative scale.
This is an automated message from the Librarian Bot. I found the following papers similar to this paper.
The following papers were recommended by the Semantic Scholar API
- ATANT v1.1: Positioning Continuity Evaluation Against Memory, Long-Context, and Agentic-Memory Benchmarks (2026)
- MemMachine: A Ground-Truth-Preserving Memory System for Personalized AI Agents (2026)
- Anatomy of Agentic Memory: Taxonomy and Empirical Analysis of Evaluation and System Limitations (2026)
- The Missing Knowledge Layer in Cognitive Architectures for AI Agents (2026)
- StatePlane: A Cognitive State Plane for Long-Horizon AI Systems Under Bounded Context (2026)
- Facts as First Class Objects: Knowledge Objects for Persistent LLM Memory (2026)
- Beyond the Context Window: A Cost-Performance Analysis of Fact-Based Memory vs. Long-Context LLMs for Persistent Agents (2026)
Please give a thumbs up to this comment if you found it helpful!
If you want recommendations for any Paper on Hugging Face checkout this Space
You can directly ask Librarian Bot for paper recommendations by tagging it in a comment: @librarian-bot recommend
Get this paper in your agent:
hf papers read 2604.06710 Don't have the latest CLI?
curl -LsSf https://hf.co/cli/install.sh | bash Models citing this paper 0
No model linking this paper
Datasets citing this paper 1
Spaces citing this paper 0
No Space linking this paper