Anthropic will let its managed agents dream

A stylized illustration of a human head in profile with an open brain filled with colorful abstract shapes, mounted on a database stack, surrounded by concentric ripple lines — visualizing the concept of AI memory consolidation and "dreaming."

In April, Anthropic launched the public beta of Managed Agents, its platform for running AI agents on its infrastructure.

On Wednesday, the company announced it is expanding Managed Agents with the ability to dream, focus on outcomes, and orchestrate multiple agents. These new features, the company says, will “make agents more capable at handling complex tasks with minimal steering.”

What do AI agents dream about?

The marquee feature here is “dreaming,” which is now in research preview. While you may not think of sleeping as a feature that will make agents more capable, the idea here is that, just like the human brain updates its memory network while a person sleeps, Claude in Managed Agents will now run a scheduled process that allows it to process and review any recent work it did. It will review recent sessions, look for patterns, and then store those updated observations in its memory.

There are controls to run this as a completely automated process, or to have users review any changes before Claude writes them into its memory.

Anthropic will let its managed agents dream
Credit: Anthropic

The practical advantage here is that by taking a more holistic look at all recent workflows — and the mistakes the agent may have made — we can find patterns an individual agent isn’t likely to see.

“Together, memory and dreaming form a robust memory system for self-improving agents.”
—Anthropic

“Together, memory and dreaming form a robust memory system for self-improving agents,” Anthropc writes. “Memory lets each agent capture what it learns as it works.”

Outcomes

The other major new feature is outcomes. The idea here is to focus on the agent’s actual intent.

“Agents do their best work when they know what ‘good’ looks like,” Anthropic says in its announcement.

To do so, they need to know the criteria for what “good” means for a given task, so users can create those for the agent, and then a separate grader agent will evaluate this output against them (that grader has its own context window, so there’s no cheating).

“Agents do their best work when they know what ‘good’ looks like.”
—Anthropic

Anthropic notes that this is especially useful for having the agent work on tasks “that require attention to detail and exhaustive coverage,” but also where there is a more subjective quality to the result (like following a brand’s voice in marketing copy).

In the company’s own testing, using outcomes improved task success by up to 10 points compared to a standard prompting loop.

Anthropic will let its managed agents dream
Credit: The New Stack

Multi-agent orchestration

One area many AI labs are currently working on is orchestrating multiple agents to work in parallel. It’s no surprise then that Anthropic, too, is bringing this to its managed agent platform. Managed Agents can now break down tasks and have a lead agent assign them to subagents.

That’s something Claude Code and Cowork often do by default, but they don’t have all that many ways to actually manage them. In Managed Agents, users now have access to an area in the Claude Console where they can see exactly what each agent did, step by step.

Availability

Both outcomes and multi-agent orchestration are now part of the public beta of Managed Agents. Users who want access to dreaming can request it here.

The post Anthropic will let its managed agents dream