
Until early 2026, building persistent memory into a Claude-powered application meant rolling your own infrastructure: a vector database, an embedding pipeline, a retrieval layer, and a chunk of glue code to stitch it all into the prompt. That has changed quickly. Anthropic shipped a native memory tool for the API, launched a managed agent runtime with memory baked in, and in early May added dreaming, multiagent orchestration, outcomes, and webhooks. A layer of third-party providers has formed alongside the API to handle the same problem from different angles.
Here is a current map of the leading options developers are using to give Claude memory that survives between sessions.
The Memory Tool. Available on the Claude API, the memory tool exposes a filesystem-style interface where Claude can create, read, update, and delete files inside a /memories directory it controls. Crucially, it runs client-side: Claude issues tool calls, the application executes them locally, and the developer chooses the storage backend — disk, S3, a database, anything. It is the lowest-level option and the most flexible, and it pairs naturally with Claude Opus 4.7 and Sonnet 4.6. The tradeoff is that the developer is responsible for path validation, retention, and any auditing.
Memory for Claude Managed Agents. Anthropic's Managed Agents runtime moved into public beta in April 2026 under the managed-agents-2026-04-01 beta header, and memory is its headline feature. Memories are stored as files on a managed filesystem with API control, audit logs, exportable stores, and rollback. Early adopters cited by Anthropic include Netflix, Rakuten, Wisedocs, and Ando, with Rakuten reporting 97% fewer first-pass errors, 27% lower cost, and 34% lower latency on long-running task agents. For production agents that need cross-session learning without standing up infrastructure, this has quickly become the default path.
On May 7, Anthropic added three capabilities on top of Managed Agents that change how memory behaves in production.
Dreaming, in research preview, is a scheduled background process that reviews past sessions, extracts patterns, and curates the memory store so agents self-improve over time. Developers choose whether dreaming updates memory automatically or stages changes for review — a key control for regulated workloads.
Multiagent orchestration, in public beta, lets a lead agent break a job into pieces and delegate to specialists with their own models, prompts, and tools, all sharing a common filesystem and memory. Outcomes and webhooks complete the picture, letting external systems subscribe to session and vault lifecycle events for observability and downstream automation.
Mem0. An open-source memory layer that has become a popular drop-in for Claude, OpenAI, and other LLMs. Mem0 ingests conversation turns, extracts what is salient, embeds it, and serves the relevant slice back at query time. It supports multiple vector databases as backends and ships with a hosted tier. For teams running multi-LLM stacks or that want a memory abstraction without committing to Anthropic's runtime, it is one of the most widely deployed third-party options.
Zep. A long-term memory service built for production agents. Zep stores chat history, summarises older turns, and exposes a temporal knowledge graph alongside vector retrieval. It works with Claude through standard tool calling and is often picked when developers need structured memory — entities, relationships, time — rather than a flat embedding store.
LangGraph and LlamaIndex memory modules. Both major orchestration frameworks now ship memory components that work with Claude out of the box. LangGraph's checkpointer and store APIs persist agent state across runs; LlamaIndex offers a memory module that handles short-term buffers and long-term vector retrieval. These are libraries rather than standalone services, but for teams already using either framework they are the path of least friction.
If the application is a chat assistant on the API and the team wants full control of storage and retention, the official memory tool is the right starting point. If it is an agent that needs to learn across sessions and the team would rather not run infrastructure, Managed Agents memory — now extended with dreaming and multiagent orchestration — is the strongest option, with measurable production results behind it. For everything in between — multi-LLM stacks, structured memory, framework-native pipelines — Mem0, Zep, and the LangGraph or LlamaIndex modules cover most of the territory.
The shape of the market in mid-2026 is clear: persistent memory has moved from a custom build problem to a stack-selection problem. The remaining work is matching the right layer to the workload — and accepting that, as with vector databases two years ago, the field will keep consolidating.