Blog

Why AI agents forget (and how to fix it)

Jonathan Bree


Forgetting is not a flaw in most AI systems. It is how they were designed. Understanding why helps explain what needs to change.

AI agents are getting remarkably capable. They can browse the web, write and execute code, manage files, send emails, and coordinate complex multi-step tasks. What most of them cannot do is remember that they did any of it. Ask the same agent the same question tomorrow and it will approach it as if for the first time.

This is not an accident. It is an architecture.


How AI Memory Actually Works Right Now

Most AI agents operate within a context window. Think of it as a whiteboard. Everything relevant to the current task gets written on it: the conversation history, any documents provided, instructions, tool outputs. The agent works from what is on the whiteboard. When the session ends, the whiteboard is wiped.

Context windows have grown substantially. Some models can now hold hundreds of thousands of tokens, enough to process entire codebases or document libraries in a single session. This is useful. It is not the same as memory.

A large context window is still a whiteboard. It is just a bigger one. When the session ends, everything goes. There is no persistence, no accumulation, no record that any of it happened.


The Three Things Missing

Genuine memory for AI agents requires three things that context windows do not provide.

The first is persistence. Information needs to survive the end of a session. Not all of it, but the right parts. Decisions made, preferences expressed, facts established, patterns observed. These should be available next time without the user having to provide them again.

The second is structure. Raw conversational history is not memory. It is a transcript. Useful memory is organised, indexed, and retrievable in ways that make it possible to surface the right thing at the right moment. This requires deliberate architecture, not just storage.

The third is relevance judgement. A system that remembers everything indiscriminately is almost as unhelpful as one that remembers nothing. Good memory involves knowing what matters, what has expired, and what is worth surfacing in a given context. This is harder than it sounds and most current systems do not attempt it.


Why This Has Been Acceptable Until Now

For simple, single-turn tasks, stateless AI works fine. Ask a question, get an answer, move on. The lack of memory is invisible because continuity was never the point.

As AI agents take on more complex, ongoing, multi-session work, the absence of memory becomes impossible to ignore. The agent that helped you draft a strategy document last week has no idea it did so. The one coordinating your workflows does not remember the constraints you established. Every session is a briefing from scratch.

This is the ceiling that stateless architecture imposes. And it is a low one.


How to Fix It

The solution is not to make context windows larger, though that helps at the margins. It is to build memory systems that sit alongside the agent and persist across sessions.

This means capturing the right information at the end of each session, storing it in a structured and retrievable form, and making it available at the start of the next one. It means building relevance judgement into the retrieval layer so that what surfaces is useful rather than merely recent. It means treating memory as a first-class component of agent architecture rather than an afterthought.

None of this is theoretically difficult. It is an engineering problem with known approaches. Vector databases, episodic memory stores, semantic indexing, retrieval-augmented generation. The pieces exist. What has been missing is a system that puts them together deliberately, with memory as the organising principle rather than a peripheral feature.


What Exabase Does Differently

Exabase builds memory into AI agents from the ground up. Persistent storage that survives sessions. Structured retrieval that surfaces context when it is relevant. A system designed around the idea that an agent should get more useful over time, not plateau at whatever it could do on day one.

The forgetting problem is solvable. Exabase is solving it.

Ship your first app in minutes.