The Clash of Multi-Agent Architectures: Cognition vs Anthropic

On June 12, Walden Yan from Cognition just published an article titled “Don’t Build Multi-Agents”. The next day, Anthropic released a lengthy 6000-word practical summary titled “How we built our multi-agent research system”.

πŸ’₯ Cognition’s Stance: Avoid Multi-Agents

The core viewpoint of Cognition co-founder Walden Yan:

The Clash of Multi-Agent Architectures: Cognition vs Anthropic

  • Context fragmentation is the root cause of agent failure;
  • Multiple agents can lead to misunderstanding tasks, inconsistent styles, and conflicting decisions;

The most stable architecture is: Single agent + linear context + precise memory compression;The Clash of Multi-Agent Architectures: Cognition vs Anthropic

β€œEven with smart agents, once you break apart the context, everything breaks.” β€”β€” Walden Yan

πŸ§ͺ Anthropic’s Counterattack: We Built Multi-Agents and Did It Well

Anthropic’s Claude Research system is a typical multi-agent parallel architecture:

The Clash of Multi-Agent Architectures: Cognition vs Anthropic

  • A Lead Agent is responsible for planning research tasks;
  • Simultaneously scheduling multiple Subagents to concurrently call tools for information search;
  • Finally, a Citation Agent adds references to generate a complete report.

They found in practical evaluations that:

🎯 The combination of Claude Opus + Sonnet multi-agents is 90.2% more accurate than a single Claude Opus.

Especially suitable for complex open-ended tasks, such as:

  • β€œIdentify the board members of all IT companies in the S&P 500”
  • β€œConduct a horizontal survey of AI agent company development paths”
  • β€œBuild an AI medical decision support map”

πŸ” Cognition’s Critique vs Anthropic’s Solutions

Cognition’s Concerns Anthropic’s Practical Engineering Solutions
❌ Subagents easily misunderstand tasks βœ… Orchestrator ensures precise division of labor and clear instruction formats
❌ Conflicting decisions lead to merged errors βœ… No shared context; the main agent is responsible for integration
❌ Huge token consumption βœ… Achieving 90% accuracy for high-value tasks is reasonable
❌ Prompts are unstable and debugging is difficult βœ… Claude self-optimizes prompts and iterates tool descriptions
❌ Difficult to debug and deploy βœ… Full-link tracing + rainbow deployment + checkpoint recovery

🧩 The Key to Anthropic’s Successful Architecture

The Clash of Multi-Agent Architectures: Cognition vs Anthropic

  • 🧠 Lead Agent plans tasks β†’ Subagents search concurrently β†’ Citation Agent adds references
  • πŸ“¦ Prompt engineering experience: Must include task objectives, output formats, tools used, and task boundaries
  • πŸ“š Built-in “heuristics” in prompts: Broad to narrow, search then compress, avoid repetition and excessive searching
  • πŸ”§ Claude can act as its own prompt engineer: Can rewrite prompts, optimize failure paths, and reduce tool usage errors
  • πŸš€ Concurrently calling 3-5 Subagents + each calling 3 tools β†’ 90% speedup

πŸ“Š Which Architecture is Right for You?

Type Cognition Approach Anthropic Approach
Programming Tasks βœ… Stable, traceable ❌ Subtask parallelism is not obvious
Forms, RPA Workflows βœ… A single agent with tools is sufficient ❌ Multiple agents are unnecessary
Research Tasks ❌ Linear agent pressure is high βœ… Multi-agent + concurrency is optimal
Enterprise Knowledge Search ❌ Single-threaded is slow βœ… Scalable, handles more context
User Interaction Agents βœ… Coherent context ❌ Multiple agents struggle to unify dialogue style

Leave a Comment