多视角辩论

debate-research

by caius-kong

Multi-perspective structured debate for complex topics. Spawn parallel subagents with opposing stances, cross-inject arguments for rebuttal, then synthesize via neutral judge into a consensus report with recommendations and scenario matrix. Use when: (1) user asks for deep comparison, pros/cons, or X vs Y analysis, (2) user asks for multi-angle research on a controversial or complex topic, (3) user explicitly requests debate, dialectical analysis, or adversarial research. NOT for: simple factual lookups, single-perspective deep research (use academic-deep-research), or quick opinion questions.

3.9k搜索与获取未扫描2026年4月6日

安装

claude skill add --url https://github.com/openclaw/skills

文档

Debate Research

Input Parameters

Collect from user before starting. Only topic is required; all others have defaults.

ParamRequiredDefaultDescription
topicyesDebate subject
rolesnoProponent + Opponent2-4 role objects: {name, stance, model?}. Default: Proponent (argue for) and Opponent (argue against). Model inherits from global.
goalnoinferredWhat question to answer
audienceno"self"Who reads the report: self / team / public
decision_typeno"personal-choice"personal-choice / team-standardization / market-analysis
evidence_roundno"auto"false / true / auto (enable when topic is fact-dense)
confirm_plannotrueShow plan and wait for user OK before execution
modelnoinheritGlobal subagent model; role-level override takes priority
output_pathnonullFile path for report; null = return in conversation

Implicit parameter: language — inferred from the user's topic/conversation language. All subagent prompts output in this language.

Example User Prompt

  • Claude Code vs OpenCode (gpt-5.4, claude-4.6-sonnet)

Execution Pipeline

Phase 0 — Pre-flight

Step 0a: Model reachability check

Collect all unique models (global + per-role + judge). For each unique model, probe via sessions_spawn with a minimal one-sentence task (e.g. "Reply OK") and model: <target>. Do NOT use curl or external HTTP — all models route through OpenClaw's provider config.

If any probe fails:

  • If user explicitly specified the failed model → abort, report failure, suggest alternatives
  • If model was default-assigned → warn user, fall back to session default model, continue

Step 0b: Plan presentation (if confirm_plan: true)

Present to user:

  • Topic
  • Role × model assignment table
  • Evidence round: on/off/auto (with rationale if auto)
  • Estimated subagent call count
  • Goal / audience / decision_type interpretation

[STOP — wait for user confirmation]

If confirm_plan: false, skip directly to Phase 1.

Phase 1 — Stance Investigation (parallel)

Spawn one subagent per role, all in parallel.

Each agent receives a prompt built from:

  • Role name + stance
  • Topic
  • web_search: enabled

Required output format per agent:

code
Core arguments (3-5):
  - [argument] | confidence: 0.0-1.0 | source: [official-docs/community-feedback/personal-blog/academic-paper]
Opponent weaknesses (2-3)
Predicted counter-attacks (1-2)

Use sessions_spawn + sessions_yield to wait for all completions.

Error handling:

  • Agent timeout → mark output [INCOMPLETE], continue pipeline

Phase 2 — Cross Rebuttal (parallel)

Spawn one subagent per role, all in parallel.

Each agent receives:

  • Its original stance
  • All other roles' Phase 1 output (cross-injected)
  • web_search: disabled

Required output format per agent:

code
Rebuttals (one per opponent argument):
  - [rebuttal] | confidence: 0.0-1.0
Weakest premise attack:
  - Identify opponent's single weakest assumption and challenge it  ← Socratic element
New attacks (2):
  - [attack]

Word limit: 300 × number_of_opponents words per agent.

Error handling:

  • Agent timeout → mark [INCOMPLETE], continue

Phase 2.5 — Evidence Audit (optional)

Triggered when evidence_round: true, or when auto and topic involves measurable claims. Auto-enable heuristic: topic contains performance benchmarks, cost comparisons, security assessments, market data, or quantitative metrics. When in doubt with auto, skip (false positive costs more than false negative).

Spawn 1 subagent as "evidence auditor":

  • Input: all Phase 1 + Phase 2 output
  • web_search: disabled
  • Task: extract every factual claim, tag each as: [official-docs] [community-feedback] [personal-blog] [no-source] [exaggerated]
  • Output: concise fact checklist

Phase 3 — Neutral Judgment

Spawn 1 subagent as neutral judge:

  • Input: Phase 1 + Phase 2 + Phase 2.5 (if available)
  • web_search: disabled
  • Weigh arguments by confidence scores AND source quality tags

Required output structure:

  1. Strong arguments per side
  2. Exaggerated claims per side
  3. Shared limitations (problems neither option solves)
  4. Core disagreements (value-level, not just factual)
  5. Consensus points
  6. Recommendation — explicit directional advice, adapted to decision_type
  7. Open Questions — unresolved unknowns that could change the conclusion
  8. Scenario selection matrix (table: scenario × recommendation × rationale)
  9. One-sentence summary

Phase 4 — Report Assembly

Orchestrator (main conversation) assembles all outputs into Markdown:

markdown
# [topic]: Debate Research Report

> **Date**: YYYY-MM-DD
> **Method**: Multi-agent structured debate (debate-research skill)
> **Roles**: [role1 (model)] | [role2 (model)] | ...
> **Audience**: [audience] | **Decision type**: [decision_type]
> **Completion**: [success | degraded-success | aborted]

## Core Arguments by Side
[Phase 1 output, organized by role]

## Cross Rebuttals
[Phase 2 output, organized by role]

## Evidence Audit
[Phase 2.5 output, or "Not requested"]

## Neutral Judgment
[Phase 3 sections 1-5]

## Recommendation
[Phase 3 section 6]

## Open Questions
[Phase 3 section 7]

## Scenario Matrix
[Phase 3 section 8]

> **One-line summary**: [Phase 3 section 9]

If output_path specified → write file. Otherwise → return in conversation.

Completion States

StateConditionBehavior
successAll phases completed normallyFull report
degraded-success1+ agents timed out or returned [INCOMPLETE]Report with degradation note
abortedModel pre-check failed / user cancelled planNo report; return error summary

Prompt Templates

See references/prompts.md for the exact prompt templates used in each phase. Orchestrator builds prompts dynamically from parameters + these templates.

相关 Skills

SEO审计工具

by amdf01-debug

热门

搜索与获取
未扫描3.9k

浏览器自动化

by chulla-ceja

热门

Browser automation CLI for AI agents. Use when the user needs to interact with websites, including navigating pages, filling forms, clicking buttons, taking screenshots, extracting data, testing web apps, or automating any browser task. Triggers include requests to "open a website", "fill out a form", "click a button", "take a screenshot", "scrape data from a page", "test this web app", "login to a site", "automate browser actions", or any task requiring programmatic web interaction.

搜索与获取
未扫描3.9k

接口规范

by alexxxiong

热门

API 规范管理工具 - 跨项目 API 文档的初始化、更新、查询与搜索。Triggers: 'API文档', 'API规范', '接口文档', '路由解析', 'apispec', 'API lookup', 'API search'.

搜索与获取
未扫描3.9k

相关 MCP 服务

by Anthropic

热门

Puppeteer 是让 Claude 自动操作浏览器进行网页抓取和测试的 MCP 服务器。

这个服务器解决了手动编写 Puppeteer 脚本的繁琐问题,适合需要自动化网页交互的开发者,比如抓取动态内容或做端到端测试。不过,作为参考实现,它可能缺少生产级的安全防护,建议在可控环境中使用。

搜索与获取
83.1k

网页抓取

编辑精选

by Anthropic

热门

Fetch 是 MCP 官方参考服务器,让 AI 能抓取网页并转为 Markdown 格式。

这个服务器解决了 AI 直接处理网页内容时格式混乱的问题,适合需要让 Claude 分析在线文档或新闻的开发者。不过作为参考实现,它缺乏生产级的安全配置,你得自己处理反爬虫和隐私风险。

搜索与获取
83.1k

Brave 搜索

编辑精选

by Anthropic

热门

Brave Search 是让 Claude 直接调用 Brave 搜索 API 获取实时网络信息的 MCP 服务器。

如果你想让 AI 助手帮你搜索最新资讯或技术文档,这个工具能绕过传统搜索的限制,直接返回结构化数据。特别适合需要实时信息的开发者,比如查 API 更新或竞品动态。不过它依赖 Brave 的 API 配额,高频使用可能受限。

搜索与获取
83.1k

评论