Beyond RAG: What Deterministic Reconstruction Actually Looks Like.
The industry is optimizing the read path: better retrieval, smarter search, fancier embeddings. The actual fix is on the write path.
RAG, knowledge graphs, and long context windows all try to solve the memory problem at read time. They search for relevant information after the fact. Deterministic reconstruction takes a different approach: decompose interactions into structured traces when they happen, so the current state can be assembled later without search. Probabilistic retrieval vs. deterministic assembly. The architecture is called DTCM.
The AI memory landscape in 2026 is crowded. Mem0, Zep, LangChain's memory modules, knowledge graph approaches, vector databases, hybrid RAG. Dozens of teams working on how to give AI systems persistent memory.
They're all working on the same problem. Most are optimizing the same side of it.
What Are the Current Approaches to AI Memory?
There are four main architectures being deployed or researched:
-
Vector RAG. Embed text as vectors, retrieve similar chunks at query time. This is the baseline. It still hallucinates 17-33% in production legal tools. Good for "find something similar." Bad for "what's the current state."
-
Knowledge Graphs. Map entities and their relationships as nodes and edges. Increasingly used alongside RAG for structural precision. Graphs offer depth while vectors offer breadth. Better at relationship queries ("who is connected to whom"). Still no temporal awareness or active/resolved tracking.
-
Hybrid RAG. Combine vectors and graphs. The 2026 consensus architecture: use vectors for breadth and graphs for depth. Incrementally better. Still fundamentally read-path. Structuring happens at query time, not storage time.
-
Agentic/Contextual Memory. Emerging category where agents maintain their own memory across tasks. Mem0 combines vector search with graph relationships and LLM-powered fact extraction. It's the closest production system to structured memory, but still retrieval-first at its core. Zep uses a temporal knowledge graph architecture that tracks entity changes over time, the only major framework with native temporal awareness, scoring 63.8% on LongMemEval vs. Mem0's 49.0%. Both are moving in the right direction. Neither decomposes at write time.
Each of these improves on the last. None of them solve the core problem.
What's the Core Problem They All Share?
Every current approach optimizes the read path: how to find the right information at query time.
The assumption: store data however it comes in (raw text, embeddings, entity-relationship pairs), then get increasingly sophisticated about retrieving it later.
This assumption breaks down for any task that requires knowing the current state of a situation.
Read-path systems answer: "What stored information is relevant to this query?" The actual need is: "What is the current living state of this situation, including what changed, what's still active, what was superseded, and what matters now?"
Those are fundamentally different operations. The first is search. The second is reconstruction. Search can approximate reconstruction sometimes, but it can't guarantee it. The data wasn't stored in a form that supports it.
What Does Write-Path-First Mean?
Write-path-first means the heavy work happens when information arrives, not when it's queried.
When a user says: "My interview at Google moved from Tuesday to Thursday. I'm less nervous now because I did two mock interviews."
A read-path system stores this as text (or embeds it as a vector). Later, retrieval might find it. Or it might find the earlier message that said "Tuesday." Or both, with no way to know which is current.
A write-path-first system decomposes this at storage time:
- Entity: Google interview
- Temporal update: Tuesday to Thursday (old state superseded)
- Emotional update: nervous to less nervous (state changed, reason: mock interviews)
- Event: two mock interviews completed (new episodic trace)
- Status: interview still active, preparation ongoing
These aren't chunks of text. They're structured traces with explicit types, timestamps, and active/resolved status. When the system needs to reconstruct this person's situation later, it doesn't search for similar text. It assembles the current state from these traces deterministically.
The data was structured for reconstruction at the moment it was stored. That's the difference.
How Does Deterministic Reconstruction Work?
At read time, the system:
- Identifies the relevant scope (which user, which context, which time range)
- Gathers all active traces within that scope
- Resolves superseded traces (Thursday replaces Tuesday, "less nervous" replaces "nervous")
- Assembles the current state from the remaining active traces
- Returns a structured situation, not a bag of chunks
There's no embedding similarity involved. No "closest vector" guessing. No "lost in the middle" degradation. Either the trace exists or it doesn't. Either it's active or it's been superseded.
| Vector RAG | Knowledge Graph | Hybrid RAG | Deterministic Reconstruction (DTCM) | |
|---|---|---|---|---|
| When structuring happens | Query time | Index time (entities) | Both | Write time (decomposition into traces) |
| What it returns | Similar chunks | Related entities | Both | Current living state |
| Handles updates | Old and new coexist | Manual edge updates | Inconsistent | Old state superseded automatically |
| Temporal awareness | None | Limited | Limited | Full: active vs. resolved, sequenced |
| Disambiguation | No scoping | Entity-based | Partial | Trace-scoped per user/context |
| Failure mode | Returns wrong chunk (silent) | Missing entity (silent) | Mixed | Returns nothing if trace doesn't exist (explicit) |
| Scales with | Embedding quality + retriever sophistication | Graph schema + query complexity | Both | Trace coverage at write time |
What Are the Five Traces?
DTCM (Decomposed Trace Convergence Memory) decomposes every interaction into five structured trace types:
- Episodic: what happened. Events, actions, conversations. The factual record.
- Emotional: how the user felt. Emotional state at the time of the interaction.
- Temporal: when it happened. Sequence, duration, deadlines, what's still active.
- Relational: who was involved. Entities, relationships, roles.
- Schematic: what pattern it fits. Recurring themes, categories, life domains.
A single sentence like "My sister Mia got the job in London and I'm really happy for her" produces traces across all five dimensions:
- Episodic: Mia got a job in London
- Emotional: User is happy
- Temporal: This is new (state change from previous "interviewing" status)
- Relational: Mia = sister, Mia to London, Mia to new job
- Schematic: Family, career milestones
These traces are stored independently. At read time, the convergence gate selects which traces are relevant to the current query and assembles them into a reconstructed situation.
This is what "beyond RAG" actually means. Not better retrieval. A different architecture entirely.
How Was This Tested?
ATANT (Automated Test for Acceptance of Narrative Truth) is the first open evaluation framework for AI continuity.
ATANT tests whether a system can maintain correct, disambiguated, updateable context across:
- 250 narrative stories spanning 6 life domains
- 1,835 verification questions
- Isolated mode: each story tested independently
- Cumulative mode: all 250 stories coexisting in the same system
Results:
- Isolated (250 stories): 1,835/1,835 correct, 100%
- Cumulative (50 stories): 304/304 correct, 100%
- Cumulative (250 stories): 1,761/1,835 correct, 96%
The cumulative result is the headline. 250 different people's lives in one system. The right fact retrieved for the right person. No cross-contamination. No hallucination from wrong-user retrieval.
No RAG system has published comparable results on a standardized continuity benchmark. No standardized continuity benchmark existed until ATANT.
What We Built
At Kenotic Labs, DTCM is the architecture. ATANT is the benchmark. The research paper is published on arXiv. The patents are filed. The reference implementation is built.
Beyond RAG means better storage, not better search. Deterministic reconstruction, not probabilistic retrieval.
Follow the research at kenoticlabs.com
Samuel Tanguturi is the founder of Kenotic Labs, building the continuity layer for AI systems. ATANT v1.0 is available on GitHub.