KS
Killer-Skills

openspec-verify-change — how to use openspec-verify-change how to use openspec-verify-change, openspec-verify-change setup guide, openspec-verify-change alternative, openspec-verify-change vs manual validation, openspec-verify-change install, what is openspec-verify-change, openspec-verify-change implementation verification

v1.0
GitHub

About this Skill

Perfect for Development Agents needing precise implementation verification against change artifacts. openspec-verify-change is a skill that verifies an implementation matches the change artifacts, such as specs, tasks, and design, using openspec list --json and AskUserQuestion tool.

Features

Runs openspec list --json to retrieve available changes
Utilizes AskUserQuestion tool for user selection of changes
Prompts for selection when no change name is provided
Displays changes with implementation tasks for efficient validation
Supports inference of change name from conversation context

# Core Topics

jixoai jixoai
[34]
[6]
Updated: 2/28/2026

Quality Score

Top 5%
55
Excellent
Based on code quality & docs
Installation
SYS Universal Install (Auto-Detect)
Cursor IDE Windsurf IDE VS Code IDE
> npx killer-skills add jixoai/openspecui/openspec-verify-change

Agent Capability Analysis

The openspec-verify-change MCP Server by jixoai is an open-source Categories.community integration for Claude and other AI agents, enabling seamless task automation and capability expansion. Optimized for how to use openspec-verify-change, openspec-verify-change setup guide, openspec-verify-change alternative.

Ideal Agent Persona

Perfect for Development Agents needing precise implementation verification against change artifacts.

Core Value

Empowers agents to ensure consistency and accuracy in projects by verifying implementation matches against specs, tasks, and design using the openspec-verify-change skill, which leverages the AskUserQuestion tool and openspec list --json command.

Capabilities Granted for openspec-verify-change MCP Server

Verifying implementation against change artifacts
Automating consistency checks for development projects
Debugging mismatches between specs and implementation

! Prerequisites & Limits

  • Requires openspec list --json command accessibility
  • Needs inference or user input for change name
  • Limited to projects with available change artifacts
Project
SKILL.md
6.1 KB
.cursorrules
1.2 KB
package.json
240 B
Ready
UTF-8

# Tags

[No tags]
SKILL.md
Readonly

Verify that an implementation matches the change artifacts (specs, tasks, design).

Input: Optionally specify a change name. If omitted, check if it can be inferred from conversation context. If vague or ambiguous you MUST prompt for available changes.

Steps

  1. If no change name provided, prompt for selection

    Run openspec list --json to get available changes. Use the AskUserQuestion tool to let the user select.

    Show changes that have implementation tasks (tasks artifact exists). Include the schema used for each change if available. Mark changes with incomplete tasks as "(In Progress)".

    IMPORTANT: Do NOT guess or auto-select a change. Always let the user choose.

  2. Check status to understand the schema

    bash
    1openspec status --change "<name>" --json

    Parse the JSON to understand:

    • schemaName: The workflow being used (e.g., "spec-driven")
    • Which artifacts exist for this change
  3. Get the change directory and load artifacts

    bash
    1openspec instructions apply --change "<name>" --json

    This returns the change directory and context files. Read all available artifacts from contextFiles.

  4. Initialize verification report structure

    Create a report structure with three dimensions:

    • Completeness: Track tasks and spec coverage
    • Correctness: Track requirement implementation and scenario coverage
    • Coherence: Track design adherence and pattern consistency

    Each dimension can have CRITICAL, WARNING, or SUGGESTION issues.

  5. Verify Completeness

    Task Completion:

    • If tasks.md exists in contextFiles, read it
    • Parse checkboxes: - [ ] (incomplete) vs - [x] (complete)
    • Count complete vs total tasks
    • If incomplete tasks exist:
      • Add CRITICAL issue for each incomplete task
      • Recommendation: "Complete task: <description>" or "Mark as done if already implemented"

    Spec Coverage:

    • If delta specs exist in openspec/changes/<name>/specs/:
      • Extract all requirements (marked with "### Requirement:")
      • For each requirement:
        • Search codebase for keywords related to the requirement
        • Assess if implementation likely exists
      • If requirements appear unimplemented:
        • Add CRITICAL issue: "Requirement not found: <requirement name>"
        • Recommendation: "Implement requirement X: <description>"
  6. Verify Correctness

    Requirement Implementation Mapping:

    • For each requirement from delta specs:
      • Search codebase for implementation evidence
      • If found, note file paths and line ranges
      • Assess if implementation matches requirement intent
      • If divergence detected:
        • Add WARNING: "Implementation may diverge from spec: <details>"
        • Recommendation: "Review <file>:<lines> against requirement X"

    Scenario Coverage:

    • For each scenario in delta specs (marked with "#### Scenario:"):
      • Check if conditions are handled in code
      • Check if tests exist covering the scenario
      • If scenario appears uncovered:
        • Add WARNING: "Scenario not covered: <scenario name>"
        • Recommendation: "Add test or implementation for scenario: <description>"
  7. Verify Coherence

    Design Adherence:

    • If design.md exists in contextFiles:
      • Extract key decisions (look for sections like "Decision:", "Approach:", "Architecture:")
      • Verify implementation follows those decisions
      • If contradiction detected:
        • Add WARNING: "Design decision not followed: <decision>"
        • Recommendation: "Update implementation or revise design.md to match reality"
    • If no design.md: Skip design adherence check, note "No design.md to verify against"

    Code Pattern Consistency:

    • Review new code for consistency with project patterns
    • Check file naming, directory structure, coding style
    • If significant deviations found:
      • Add SUGGESTION: "Code pattern deviation: <details>"
      • Recommendation: "Consider following project pattern: <example>"
  8. Generate Verification Report

    Summary Scorecard:

    ## Verification Report: <change-name>
    
    ### Summary
    | Dimension    | Status           |
    |--------------|------------------|
    | Completeness | X/Y tasks, N reqs|
    | Correctness  | M/N reqs covered |
    | Coherence    | Followed/Issues  |
    

    Issues by Priority:

    1. CRITICAL (Must fix before archive):

      • Incomplete tasks
      • Missing requirement implementations
      • Each with specific, actionable recommendation
    2. WARNING (Should fix):

      • Spec/design divergences
      • Missing scenario coverage
      • Each with specific recommendation
    3. SUGGESTION (Nice to fix):

      • Pattern inconsistencies
      • Minor improvements
      • Each with specific recommendation

    Final Assessment:

    • If CRITICAL issues: "X critical issue(s) found. Fix before archiving."
    • If only warnings: "No critical issues. Y warning(s) to consider. Ready for archive (with noted improvements)."
    • If all clear: "All checks passed. Ready for archive."

Verification Heuristics

  • Completeness: Focus on objective checklist items (checkboxes, requirements list)
  • Correctness: Use keyword search, file path analysis, reasonable inference - don't require perfect certainty
  • Coherence: Look for glaring inconsistencies, don't nitpick style
  • False Positives: When uncertain, prefer SUGGESTION over WARNING, WARNING over CRITICAL
  • Actionability: Every issue must have a specific recommendation with file/line references where applicable

Graceful Degradation

  • If only tasks.md exists: verify task completion only, skip spec/design checks
  • If tasks + specs exist: verify completeness and correctness, skip design
  • If full artifacts: verify all three dimensions
  • Always note which checks were skipped and why

Output Format

Use clear markdown with:

  • Table for summary scorecard
  • Grouped lists for issues (CRITICAL/WARNING/SUGGESTION)
  • Code references in format: file.ts:123
  • Specific, actionable recommendations
  • No vague suggestions like "consider reviewing"

Related Skills

Looking for an alternative to openspec-verify-change or building a Categories.community AI Agent? Explore these related open-source MCP Servers.

View All

widget-generator

Logo of f
f

widget-generator is an open-source AI agent skill for creating widget plugins that are injected into prompt feeds on prompts.chat. It supports two rendering modes: standard prompt widgets using default PromptCard styling and custom render widgets built as full React components.

149.6k
0
Design

chat-sdk

Logo of lobehub
lobehub

chat-sdk is a unified TypeScript SDK for building chat bots across multiple platforms, providing a single interface for deploying bot logic.

73.0k
0
Communication

zustand

Logo of lobehub
lobehub

The ultimate space for work and life — to find, build, and collaborate with agent teammates that grow with you. We are taking agent harness to the next level — enabling multi-agent collaboration, effortless agent team design, and introducing agents as the unit of work interaction.

72.8k
0
Communication

data-fetching

Logo of lobehub
lobehub

The ultimate space for work and life — to find, build, and collaborate with agent teammates that grow with you. We are taking agent harness to the next level — enabling multi-agent collaboration, effortless agent team design, and introducing agents as the unit of work interaction.

72.8k
0
Communication