Anthropic Unveils 'Dreaming' AI System That Lets Agents Learn From Their Own Mistakes
Anthropic introduces 'dreaming,' a system that enables AI agents to learn from past sessions and improve over time, along with updates to its Claude Managed Agents platform.

["Anthropic on Tuesday unveiled a suite of updates to its Claude Managed Agents platform at its second annual Code with Claude developer conference in San Francisco, introducing a new capability called 'dreaming' that lets AI agents learn from their own past sessions and improve over time — a step toward the kind of self-correcting, self-improving AI systems that enterprises have demanded before trusting agents with production workloads.", 'The company also moved two previously experimental features — outcomes and multi-agent orchestration — from research preview into public beta, making them broadly available to developers building on the Claude platform. Together, the three features address what Anthropic says are the hardest problems in running AI agents at scale: keeping them accurate, helping them learn, and preventing them from becoming bottlenecks on complex, multi-step work.', 'Early adopters are already reporting significant results. Legal AI company Harvey saw task completion rates increase roughly 6x after implementing dreaming.
Medical document review company Wisedocs cut its document review time by 50% using outcomes. And Netflix is now processing logs from hundreds of builds simultaneously using multi-agent orchestration. The announcements come at a moment of extraordinary momentum for Anthropic.', "CEO Dario Amodei disclosed during a fireside chat at the conference that the company's growth has outpaced even its own aggressive internal projections.
In the first quarter of 2026, Anthropic saw what Amodei described as 80x annualized growth in revenue and usage — far exceeding the 10x annual growth the company had planned for. API volume on the Claude platform is up nearly 70x year over year, and the average developer using Claude Code now spends 20 hours per week working with the tool.", "The dreaming feature works by reviewing an agent's past sessions and memory stores, extracting patterns across them, and curating those memories so agents improve over time. It surfaces insights that no single agent session could see on its own: recurring mistakes, workflows that multiple agents converge on independently, and preferences shared across a team of agents.", "The updates also include a live demo that showed AI agents improving overnight without human guidance, and a broader platform push that Anthropic framed throughout the conference as closing 'the gap between what AI can do and what it's actually doing for people.' The competitive implications are significant, as AI agent platforms from OpenAI, Google, and others compete for developer adoption, Anthropic is betting that production reliability — not just raw model intelligence — will determine which platform wins enterprise budgets."]
Source: VentureBeat