design-tournament — community design-tournament, scratch-scanner-rs, community, ide skills

v1.0.0

About this Skill

Perfect for Collaborative AI Agents needing parallel design exploration and evaluation capabilities. Run a parallel diverge-then-converge design tournament — 3-5 independent agents explore a problem, then 2 ranking agents evaluate and stack-rank the results with confidence scores

ahrav ahrav
[0]
[0]
Updated: 3/7/2026

Killer-Skills Review

Decision support comes first. Repository text comes second.

Reviewed Landing Page Review Score: 9/11

Killer-Skills keeps this page indexable because it adds recommendation, limitations, and review signals beyond the upstream repository text.

Original recommendation layer Concrete use-case guidance Explicit limitations and caution Quality floor passed for review Locale and body language aligned
Review Score
9/11
Quality Score
60
Canonical Locale
en
Detected Body Locale
en

Perfect for Collaborative AI Agents needing parallel design exploration and evaluation capabilities. Run a parallel diverge-then-converge design tournament — 3-5 independent agents explore a problem, then 2 ranking agents evaluate and stack-rank the results with confidence scores

Core Value

Empowers agents to facilitate independent design exploration, evaluation, and stack-ranking using a two-phase 'wisdom of crowds' pattern, leveraging confidence-weighted rankings and supporting multiple valid architectural approaches through protocols like API design and data structure development.

Ideal Agent Persona

Perfect for Collaborative AI Agents needing parallel design exploration and evaluation capabilities.

Capabilities Granted for design-tournament

Exploring multiple valid architectural approaches for complex projects
Evaluating and stack-ranking design results for new feature development
Breaking down tasks into independent design decisions for parallel exploration

! Prerequisites & Limits

  • Requires multiple independent agents for design exploration
  • Needs independent rankers for evaluation and stack-ranking
  • Limited to design decision-making and evaluation processes

Source Boundary

The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.

After The Review

Decide The Next Action Before You Keep Reading Repository Material

Killer-Skills should not stop at opening repository instructions. It should help you decide whether to install this skill, when to cross-check against trusted collections, and when to move into workflow rollout.

Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is design-tournament?

Perfect for Collaborative AI Agents needing parallel design exploration and evaluation capabilities. Run a parallel diverge-then-converge design tournament — 3-5 independent agents explore a problem, then 2 ranking agents evaluate and stack-rank the results with confidence scores

How do I install design-tournament?

Run the command: npx killer-skills add ahrav/scratch-scanner-rs/design-tournament. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for design-tournament?

Key use cases include: Exploring multiple valid architectural approaches for complex projects, Evaluating and stack-ranking design results for new feature development, Breaking down tasks into independent design decisions for parallel exploration.

Which IDEs are compatible with design-tournament?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for design-tournament?

Requires multiple independent agents for design exploration. Needs independent rankers for evaluation and stack-ranking. Limited to design decision-making and evaluation processes.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add ahrav/scratch-scanner-rs/design-tournament. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use design-tournament immediately in the current project.

Upstream Repository Material

The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.

Upstream Source

design-tournament

Install design-tournament, an AI agent skill for AI agent workflows and automation. Review the use cases, limitations, and setup path before rollout.

SKILL.md
Readonly
Upstream Repository Material
The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.
Supporting Evidence

Design Tournament

A two-phase "wisdom of crowds" pattern for design work. Multiple independent agents explore the same problem in isolation, then independent rankers evaluate and merge the findings into a confidence-weighted stack ranking.

When to Use

  • Designing a new feature, data structure, algorithm, or API
  • Exploring multiple valid architectural approaches
  • Story/task breakdown where the "right" decomposition isn't obvious
  • Any design decision where independent perspectives reduce blind spots

Invocation

design-tournament <problem statement>

The <problem statement> is the full description of what needs to be designed. If no argument is given, ask the user for the problem statement before proceeding.

Phase 1 — Diverge (Parallel Exploration)

Launch 5 independent agents in parallel using the Codex agent tools (spawn_agent, send_input, wait) with agent_type=default. Every agent receives identical instructions so that their outputs are independently generated with zero cross-contamination.

All 5 agents MUST be launched in a single message (one message, five spawn_agent calls) so they run concurrently.

Agent Prompt Template (identical for all 5)

Each agent receives this prompt, with {PROBLEM} replaced by the user's problem statement and {AGENT_ID} set to a unique label (Alpha, Beta, Gamma, Delta, Epsilon):

You are Design Agent {AGENT_ID} in an independent design tournament.

## Problem
{PROBLEM}

## Your Task

Produce a COMPLETE, standalone design proposal. You are working independently —
do not assume anyone else's output exists.

### Required Sections

1. **Problem Restatement** — Restate the problem in your own words to confirm
   understanding. Call out any ambiguities or assumptions you are making.

2. **Design Proposal** — Your recommended design. Include:
   - Core idea / approach (1-2 sentence summary)
   - Detailed description with concrete types, signatures, or pseudocode
   - Key invariants or contracts the design enforces
   - How the design handles edge cases and failure modes

3. **Alternatives Considered** — At least 2 alternative approaches you
   evaluated and why you rejected them.

4. **Trade-offs** — Explicit trade-off analysis:
   - What does this design optimize for?
   - What does it sacrifice?
   - Under what conditions would a different design be better?

5. **Risks & Open Questions** — What could go wrong? What remains uncertain?

6. **Self-Assessment** — Rate your own confidence (0-100) in this design and
   explain why.

### Rules

- Be specific and concrete — pseudocode or type signatures over hand-waving.
- Do NOT hedge with "it depends" — commit to a design and defend it.
- Explore the codebase if you need context (use `rg --files`, `rg`, and targeted file reads).
- Aim for depth over breadth. One well-defended design beats three sketches.

### Output Format

Return your response as a single markdown document with the sections above.
Start with: `# Design Proposal — Agent {AGENT_ID}`

Collecting Results

After all 5 agents complete, gather their outputs. If any agent fails or times out, proceed with the agents that succeeded (minimum 3 required).

Phase 2 — Converge (Parallel Ranking)

Launch 2 independent ranking agents in parallel using the Codex agent tools (spawn_agent, send_input, wait) with agent_type=default. Both receive identical instructions and the collected Phase 1 outputs.

Both rankers MUST be launched in a single message (one message, two spawn_agent calls) so they run concurrently.

Ranker Prompt Template (identical for both)

You are Ranking Judge {JUDGE_ID} in a design tournament.

## Original Problem
{PROBLEM}

## Design Proposals

Below are {N} independently generated design proposals for the same problem.
Evaluate them WITHOUT knowing which agent produced each one (they are labeled
Alpha through Epsilon for reference only).

{PROPOSALS}

## Your Task

### 1. Evaluation Criteria

Score each proposal on these dimensions (1-10 scale):

| Criterion | Weight | Description |
|-----------|--------|-------------|
| **Correctness** | 25% | Does the design actually solve the problem? Are invariants sound? |
| **Simplicity** | 20% | Is this the simplest design that could work? Minimal moving parts? |
| **Robustness** | 20% | How well does it handle edge cases, failures, and adversarial inputs? |
| **Extensibility** | 15% | Can the design accommodate likely future requirements without rework? |
| **Specificity** | 10% | How concrete and implementable is the proposal? |
| **Trade-off Awareness** | 10% | Does the author understand what they're sacrificing? |

### 2. Per-Proposal Evaluation

For EACH proposal, provide:
- Scores on each criterion
- Weighted total score
- Top strength (1 sentence)
- Top weakness (1 sentence)
- Key insight unique to this proposal (if any)

### 3. Stack Ranking

Produce a final ranking from best to worst. For each position:
- Rank position (1 = best)
- Agent label
- Weighted total score
- Confidence in this ranking position (0-100):
  - 90-100: Clear winner/loser, large score gap
  - 70-89: Solid ranking, meaningful score gap
  - 50-69: Close call, could reasonably swap with adjacent rank
  - Below 50: Essentially tied, ranking is arbitrary

### 4. Synthesis Recommendation

After ranking, provide:
- **Recommended design**: Which proposal to use as the starting point
- **Cherry-pick list**: Specific ideas from OTHER proposals that should be
  incorporated into the winner
- **Composite confidence** (0-100): Your overall confidence that the #1
  ranked proposal (with cherry-picks) is the right design

### Rules

- Judge designs on MERIT, not on writing quality or verbosity.
- If two designs are substantively identical, say so — don't force a
  false distinction.
- Be willing to rank a simple, correct design above an elaborate one.
- Explore the codebase if you need context to evaluate feasibility.

### Output Format

Return your response as a single markdown document.
Start with: `# Ranking Report — Judge {JUDGE_ID}`

Merging Rankings

After both rankers complete, merge their results:

  1. For each proposal, average the weighted scores from both judges.
  2. For each rank position, compute:
    • Average score
    • Agreement: did both judges assign the same rank? (yes/no)
    • Combined confidence: average of both judges' per-position confidence, reduced by 15 points if they disagree on rank position
  3. Final stack ranking by averaged weighted score (descending).
  4. Synthesize cherry-pick lists from both judges into a unified list.

Final Output Format

Present the merged result to the user as:

markdown
1## Design Tournament Results 2 3### Problem 4{one-line restatement} 5 6### Stack Ranking 7 8| Rank | Agent | Avg Score | Judge Agreement | Confidence | Key Strength | 9|------|-------|-----------|-----------------|------------|--------------| 10| 1 | ... | 8.2/10 | Yes | 92% | ... | 11| 2 | ... | 7.5/10 | Yes | 78% | ... | 12| 3 | ... | 6.8/10 | No (J1:#2,J2:#4)| 54% | ... | 13| ... | ... | ... | ... | ... | ... | 14 15### Recommended Design 16 17**Base**: Agent {winner}'s proposal 18**Cherry-picks from other proposals**: 19- From Agent X: {specific idea} 20- From Agent Y: {specific idea} 21 22**Composite Confidence**: {average of both judges' composite}% 23 24### Winning Proposal Summary 25{2-3 paragraph summary of the #1 design with cherry-picks integrated} 26 27### Full Proposals (collapsed) 28<details><summary>Agent Alpha — Score: X.X</summary> 29{full proposal text} 30</details> 31{repeat for each agent} 32 33### Ranking Details (collapsed) 34<details><summary>Judge 1 Report</summary> 35{full ranking report} 36</details> 37<details><summary>Judge 2 Report</summary> 38{full ranking report} 39</details>

Configuration

The default is 5 explorers + 2 rankers (7 total agents). If the user specifies a different count (e.g., design-tournament 3 agents: ...), respect it but enforce these minimums:

  • Phase 1: at least 3 explorer agents
  • Phase 2: always exactly 2 ranking agents

Tips

  • For large or ambiguous problems, include relevant file paths or context in the problem statement so explorer agents can read the codebase.
  • If the problem involves existing code, tell agents which files to examine.
  • The skill works for any design task: features, data structures, algorithms, API shapes, system architecture, task decomposition, migration plans.

Related Skills

Looking for an alternative to design-tournament or another community skill for your workflow? Explore these related open-source skills.

View All

openclaw-release-maintainer

Logo of openclaw
openclaw

Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞

333.8k
0
AI

widget-generator

Logo of f
f

Generate customizable widget plugins for the prompts.chat feed system

149.6k
0
AI

flags

Logo of vercel
vercel

The React Framework

138.4k
0
Browser

pr-review

Logo of pytorch
pytorch

Tensors and Dynamic neural networks in Python with strong GPU acceleration

98.6k
0
Developer