Claude Managed Agents
Anthropic Managed Agents Add Memory — Persistent State for AI That Actually Ships
Anthropic has added persistent memory stores to its Managed Agents platform, giving AI agents the ability to retain knowledge across sessions without custom infrastructure. The update turns Claude from a stateless chat model into a long‑running worker that picks up where it left off — and it changes how builders architect agentic workflows.
What Managed Agents with Memory Actually Does
Anthropic's Managed Agents (public beta, launched April 8, 2026) is a hosted infrastructure layer on the Claude Platform that runs long‑horizon AI agents. Developers define what an agent does — model, prompt, tools, MCP servers — and Anthropic handles where and how it runs: sandboxed execution, tool authentication, session state, error recovery, and orchestration.
Memory on Managed Agents (added April 23, 2026 to public beta) gives agents persistent knowledge across sessions. Each memory store is a workspace‑scoped collection of text documents mounted as a directory (/mnt/memory/) inside the agent's container. The agent reads and writes memories using the same bash and file tools it already uses, according to Anthropic's engineering blog. No new API patterns needed.
- Filesystem‑mounted memory Not a vector database or special API. Claude uses existing bash and code execution tools to read/write memory files, leveraging its already‑strong file manipulation capabilities.
- Versioned and auditable Every memory change creates an immutable version for audit and rollback. All writes appear in the session event stream for tracing.
- Multi‑agent sharing Multiple agents can share the same store concurrently without overwriting each other’s data.
- Scoped access Up to 8 memory stores per session, each capped at ~100KB (~25K tokens). Stores can be read_only (for shared reference material) or read_write.
The Architecture: Decoupling the Brain from the Hands
Anthropic decouples three components that were previously tangled in a single container, as Anthropic's engineering blog explains:
| Component | Role | Key Property |
|---|---|---|
| Session | Append‑only log of everything that happened | Durable, stored outside the orchestration loop |
| Orchestration Loop | Loop that calls Claude and routes tool calls | Stateless, can be rebooted without data loss |
| Sandbox | Execution environment for code/file edits | Interchangeable (“cattle” not “pet”) |
The key insight: Session ≠ Context Window. The session log lives outside Claude's context window. Claude can interrogate it via getEvents() — picking up from where it stopped, rewinding, or re‑reading context before a specific action. This avoids irreversible compaction decisions about what to keep.
On security: credentials never reach the sandbox. Auth is either bundled with the resource (e.g., a git token clones the repo during sandbox init) or stored in an external vault accessed via dedicated MCP proxy. The orchestration layer is never made aware of credentials.
How It Compares to OpenAI and Google
The market is split between open‑source SDK (OpenAI) vs. bundled managed service (Anthropic/Google), mirroring the Terraform vs. CloudFormation divide, according to The New Stack.
| Dimension | Anthropic Managed Agents | OpenAI Agents SDK | Google Vertex AI Agent Engine |
|---|---|---|---|
| Hosting | Fully managed by Anthropic | Self‑hosted (7 sandbox providers) | Managed runtime on Vertex |
| Pricing | Token rates + $0.08/session‑hour | Standard API pricing, no runtime fee | Separate consumption lines |
| Model lock‑in | Claude only (Opus/Sonnet/Haiku) | OpenAI‑optimized; supports 100+ LLMs | Google models |
| Memory | Filesystem‑mounted persistent stores with versioning | Externalized state via snapshotting | Separate memory consumption line |
| MCP | Deepest integration (Anthropic built MCP) | Adopted early 2026 | Google’s own connectors |
OpenAI explicitly framed managed agent APIs as "simplifying deployment at the cost of constraining where agents run and how they access sensitive data" — a direct shot at Anthropic's managed approach, per The New Stack.
The Infrastructure Problem Nobody Talks About
As Data Center Knowledge reports, quoting Sameh Boujelbene (VP at Dell'Oro Group), the performance bottleneck is shifting from GPU throughput to storage, networking, and data movement. Each request now triggers repeated internal hops between models, memory stores, tool sandboxes, and schedulers.
Multi‑step stateful workflows are far more latency‑sensitive than single inference calls — roughly 10x more sensitive. The architecture demands what Data Center Knowledge describes as "low tail‑latency lossless fabrics" with a clearer split between scale‑up (powers the brain) and scale‑out (connects brain to memory, tools, state). Networking is expected to gain ~10 percentage points in overall data center IT spend by end of the decade.
For builders, this means agent infrastructure costs won't just be about GPU tokens. Storage I/O, network latency between components, and state persistence are becoming first‑class cost factors.
What Developers Are Saying
Reaction on Hacker News is mixed but engaged. The top themes:
- Vendor lock‑in is the #1 concern. "This is a lock‑in into their SDK and their format," wrote Weilun Chen (founder) on Hacker News. Claude‑only means no GPT‑5, Gemini, or DeepSeek. Migration is non‑trivial.
- The orchestration layer quality debate. "The orchestration layer is kind of buggy. The LLM still wanders and cycles. It's a monolithic LLM herding machine. The underlying model is awesome and the orchestration layer works well enough," noted HN user steve_adams_86.
- Pricing math at scale. "If you have the engineering capacity to run your own agent infrastructure, the session costs may exceed doing it yourself," per Sathish Raju's fine‑print analysis.
- Most powerful features aren't GA yet. Multi‑agent coordination and self‑evaluation require separate access requests. "If your use case depends on autonomous multi‑agent parallelism, you are not deploying that next week."
On the positive side, InfoQ quotes Radhika Menon from NTT DATA: "All the infrastructure complexity that used to take months is now native to the platform," as InfoQ reported. Launch customers including Notion, Rakuten, Sentry, Asana, and Atlassian are already running production workloads.
"This is a lock-in into their SDK and their format. Claude-only means no GPT-5, Gemini, or DeepSeek. Migration is non-trivial."
The Builder Takeaway
Managed Agents with Memory solve a real pain: the months of infrastructure work needed before shipping an agent. Sandboxed execution, auth wiring, retry logic, state management — all pre‑built. The trade‑off is vendor lock‑in to Claude and Anthropic's platform.
For builders already in the Claude ecosystem, the value proposition is clear: from idea to production in days instead of months, at $0.08/session‑hour on top of token costs. For builders who need model flexibility, OpenAI's Agents SDK offers the open‑source path — but you carry the infrastructure burden yourself.
The memory addition is the most significant update. Persistent, versioned, filesystem‑mounted memory that survives crashes and is shared across agents addresses the single biggest complaint about AI agents: they forget everything between sessions. As SD Times notes, Opus 4.7 is specifically tuned to be better at using filesystem‑based memory — it remembers important notes across long multi‑session work and uses them to move on to new tasks requiring less upfront context.
Watch for two things next: multi‑agent coordination graduating from research preview to GA, and Anthropic's pricing model evolving as builders push session‑hour costs at scale.
Apr 28, 2026
Anthropic Valuation Hits $1 Trillion on Secondary Markets, Overtaking OpenAI
Anthropic’s implied valuation has surpassed $1 trillion on secondary markets, overtaking OpenAI for the first time. The 2.6x jump from its February Series G comes as Claude Code adoption drives revenue growth from $9 billion to over $30 billion in four months — but the numbers come from thin, illiquid trading that may not hold at IPO.
Apr 28, 2026
Amazon Court Order Blocks Perplexity AI Bots — New CFAA Precedent for AI Agents
A federal judge ordered Perplexity to stop its Comet AI shopping agent from accessing Amazon Prime accounts, establishing the first legal precedent that the CFAA applies to autonomous AI agents — even when users authorize them. The ruling forces every builder working with agentic AI to rethink how their tools access third-party platforms.
Apr 28, 2026
600 Google Employees Demand Pichai Reject Classified Pentagon AI Deal
More than 600 Google employees — including DeepMind staff and VPs — signed a letter demanding CEO Sundar Pichai refuse classified Pentagon AI work, directly referencing the Anthropic precedent where the DoD dropped Claude for requesting similar restrictions.
Related News
Apr 28, 2026
OpenAI Symphony Turns Linear Boards Into Autonomous Coding Agent Orchestration
OpenAI released Symphony, an open-source orchestration spec that turns Linear issue trackers into control planes for autonomous Codex agents. Teams report a 500% increase in landed PRs, and the framework now supports multi-model runtimes beyond Codex.
Apr 27, 2026
Claude Managed Agents Get Persistent Memory in Public Beta
Anthropic has launched persistent memory for Claude Managed Agents in public beta, enabling AI agents to learn across sessions. Early adopters like Rakuten report 97% fewer errors and 27% lower costs. Here's how the filesystem-based memory layer works and what it means for builders.