Debate + consensus

Debate is a multi-agent pattern where two or more agents argue opposing sides of a question. A judge, another agent or a human, evaluates and picks a winner or synthesizes. It sounds exotic. It actually works: on many reasoning tasks, debate produces more accurate answers than a single agent running alone. The trick is that arguing against yourself surfaces flaws your first-draft reasoning missed.

The setup

  1. The question is posed.
  2. Agent A is assigned one side (e.g., "argue yes").
  3. Agent B is assigned the opposite side, or the "find the flaws" role.
  4. They exchange arguments for N rounds (usually 2-4).
  5. A judge reads the full transcript and either picks a side or synthesizes.

Why it works

A single agent's first-pass reasoning can be confidently wrong in a consistent direction. When you force the same model to argue the opposite side, it's now actively looking for holes in the first position. The holes that exist, it finds. The judge, seeing both sides, has more signal than either agent alone had. Result: better final answers on problems where single-agent output tends to be inconsistent.

When to reach for debate

A worked example: a fact-check

Claim under review: "Startup X raised $50M Series B in March 2026."

  1. Claim agent: searches for evidence of the raise. Finds two press releases and a Crunchbase entry. Argues: "Yes, the raise happened."
  2. Skeptic agent: searches for contradictions. Notices the press releases are both from the company itself, no third-party reporting, Crunchbase entry created recently with minimal provenance. Argues: "The evidence chain is thin."
  3. Round 2: claim agent defends by finding independent coverage. Skeptic notes the coverage dates are all after the initial press release, no primary-source confirmation.
  4. Judge: "Claim is plausible but unverified. Recommend flagging as 'reported, unverified' rather than 'confirmed.'"

A single agent would likely have taken the press releases at face value. Debate caught the soft evidence chain.

Is it worth the cost?

The empirical picture

Published benchmarks show debate-based agents outperforming single-agent baselines on math, logic, fact-checking, and multi-step reasoning tasks. The gains are biggest where single-agent outputs are inconsistent run-to-run. For tasks the model nails the first time, debate is just expensive noise.

Failure modes

Prompting notes

What to do with this