On June 12, Walden Yan from Cognition just published an article titled “Donβt Build Multi-Agents”. The next day, Anthropic released a lengthy 6000-word practical summary titled “How we built our multi-agent research system”.
π₯ Cognition’s Stance: Avoid Multi-Agents
The core viewpoint of Cognition co-founder Walden Yan:

- Context fragmentation is the root cause of agent failure;
- Multiple agents can lead to misunderstanding tasks, inconsistent styles, and conflicting decisions;
The most stable architecture is: Single agent + linear context + precise memory compression;
βEven with smart agents, once you break apart the context, everything breaks.β ββ Walden Yan
π§ͺ Anthropic’s Counterattack: We Built Multi-Agents and Did It Well
Anthropic’s Claude Research system is a typical multi-agent parallel architecture:

- A Lead Agent is responsible for planning research tasks;
- Simultaneously scheduling multiple Subagents to concurrently call tools for information search;
- Finally, a Citation Agent adds references to generate a complete report.
They found in practical evaluations that:
π― The combination of Claude Opus + Sonnet multi-agents is 90.2% more accurate than a single Claude Opus.
Especially suitable for complex open-ended tasks, such as:
- βIdentify the board members of all IT companies in the S&P 500β
- βConduct a horizontal survey of AI agent company development pathsβ
- βBuild an AI medical decision support mapβ
π Cognition’s Critique vs Anthropic’s Solutions
| Cognition’s Concerns | Anthropic’s Practical Engineering Solutions |
|---|---|
| β Subagents easily misunderstand tasks | β Orchestrator ensures precise division of labor and clear instruction formats |
| β Conflicting decisions lead to merged errors | β No shared context; the main agent is responsible for integration |
| β Huge token consumption | β Achieving 90% accuracy for high-value tasks is reasonable |
| β Prompts are unstable and debugging is difficult | β Claude self-optimizes prompts and iterates tool descriptions |
| β Difficult to debug and deploy | β Full-link tracing + rainbow deployment + checkpoint recovery |
π§© The Key to Anthropic’s Successful Architecture

- π§ Lead Agent plans tasks β Subagents search concurrently β Citation Agent adds references
- π¦ Prompt engineering experience: Must include task objectives, output formats, tools used, and task boundaries
- π Built-in “heuristics” in prompts: Broad to narrow, search then compress, avoid repetition and excessive searching
- π§ Claude can act as its own prompt engineer: Can rewrite prompts, optimize failure paths, and reduce tool usage errors
- π Concurrently calling 3-5 Subagents + each calling 3 tools β 90% speedup
π Which Architecture is Right for You?
| Type | Cognition Approach | Anthropic Approach |
|---|---|---|
| Programming Tasks | β Stable, traceable | β Subtask parallelism is not obvious |
| Forms, RPA Workflows | β A single agent with tools is sufficient | β Multiple agents are unnecessary |
| Research Tasks | β Linear agent pressure is high | β Multi-agent + concurrency is optimal |
| Enterprise Knowledge Search | β Single-threaded is slow | β Scalable, handles more context |
| User Interaction Agents | β Coherent context | β Multiple agents struggle to unify dialogue style |