KS
Killer-Skills

skill-stocktake — how to use skill-stocktake how to use skill-stocktake, what is skill-stocktake, skill-stocktake vs manual audit, skill-stocktake alternative, skill-stocktake install, skill-stocktake setup guide, Claude Code skill audit, MCP developer tools, AI agent quality checklist

Verified
v1.0.0
GitHub

About this Skill

Essential for Quality Assurance Agents auditing Claude skill repositories across global and project-level directories. skill-stocktake is a slash command (/skill-stocktake) that audits Claude skills and commands for quality using a checklist and AI holistic judgment. It supports Quick Scan mode for recently changed skills and Full Stocktake mode for a complete review of all skills.

Features

Executes via the /skill-stocktake slash command
Performs AI holistic judgment for quality auditing
Supports Quick Scan mode for recently changed skills only
Supports Full Stocktake mode for a complete review
Scans global skills in ~/.claude/skills/ directory
Scans project-level skills in {cwd}/.claude/skills/ directory

# Core Topics

affaan-m affaan-m
[62.0k]
[7678]
Updated: 3/6/2026

Quality Score

Top 5%
80
Excellent
Based on code quality & docs
Installation
SYS Universal Install (Auto-Detect)
Cursor IDE Windsurf IDE VS Code IDE
> npx killer-skills add affaan-m/everything-claude-code/skill-stocktake

Agent Capability Analysis

The skill-stocktake MCP Server by affaan-m is an open-source Categories.official integration for Claude and other AI agents, enabling seamless task automation and capability expansion. Optimized for how to use skill-stocktake, what is skill-stocktake, skill-stocktake vs manual audit.

Ideal Agent Persona

Essential for Quality Assurance Agents auditing Claude skill repositories across global and project-level directories.

Core Value

Enables automated batch evaluation of Claude skills using sequential subagent processing with quality checklist validation and AI holistic judgment. Supports both Quick Scan (changed skills only) and Full Stocktake modes for comprehensive content analysis.

Capabilities Granted for skill-stocktake MCP Server

Auditing global skills in ~/.claude/skills/ directory
Validating project-level skills in {cwd}/.claude/skills/
Performing sequential subagent batch evaluation
Conducting Quick Scan reviews for recently modified skills

! Prerequisites & Limits

  • Requires access to filesystem paths
  • Limited to Claude skill ecosystem
  • Dependent on directory structure compliance
Project
SKILL.md
7.1 KB
.cursorrules
1.2 KB
package.json
240 B
Ready
UTF-8
SKILL.md
Readonly

skill-stocktake

Slash command (/skill-stocktake) that audits all Claude skills and commands using a quality checklist + AI holistic judgment. Supports two modes: Quick Scan for recently changed skills, and Full Stocktake for a complete review.

Scope

The command targets the following paths relative to the directory where it is invoked:

PathDescription
~/.claude/skills/Global skills (all projects)
{cwd}/.claude/skills/Project-level skills (if the directory exists)

At the start of Phase 1, the command explicitly lists which paths were found and scanned.

Targeting a specific project

To include project-level skills, run from that project's root directory:

bash
1cd ~/path/to/my-project 2/skill-stocktake

If the project has no .claude/skills/ directory, only global skills and commands are evaluated.

Modes

ModeTriggerDuration
Quick Scanresults.json exists (default)5–10 min
Full Stocktakeresults.json absent, or /skill-stocktake full20–30 min

Results cache: ~/.claude/skills/skill-stocktake/results.json

Quick Scan Flow

Re-evaluate only skills that have changed since the last run (5–10 min).

  1. Read ~/.claude/skills/skill-stocktake/results.json
  2. Run: bash ~/.claude/skills/skill-stocktake/scripts/quick-diff.sh \ ~/.claude/skills/skill-stocktake/results.json (Project dir is auto-detected from $PWD/.claude/skills; pass it explicitly only if needed)
  3. If output is []: report "No changes since last run." and stop
  4. Re-evaluate only those changed files using the same Phase 2 criteria
  5. Carry forward unchanged skills from previous results
  6. Output only the diff
  7. Run: bash ~/.claude/skills/skill-stocktake/scripts/save-results.sh \ ~/.claude/skills/skill-stocktake/results.json <<< "$EVAL_RESULTS"

Full Stocktake Flow

Phase 1 — Inventory

Run: bash ~/.claude/skills/skill-stocktake/scripts/scan.sh

The script enumerates skill files, extracts frontmatter, and collects UTC mtimes. Project dir is auto-detected from $PWD/.claude/skills; pass it explicitly only if needed. Present the scan summary and inventory table from the script output:

Scanning:
  ✓ ~/.claude/skills/         (17 files)
  ✗ {cwd}/.claude/skills/    (not found — global skills only)
Skill7d use30d useDescription

Phase 2 — Quality Evaluation

Launch a Task tool subagent (Explore agent, model: opus) with the full inventory and checklist. The subagent reads each skill, applies the checklist, and returns per-skill JSON:

{ "verdict": "Keep"|"Improve"|"Update"|"Retire"|"Merge into [X]", "reason": "..." }

Chunk guidance: Process ~20 skills per subagent invocation to keep context manageable. Save intermediate results to results.json (status: "in_progress") after each chunk.

After all skills are evaluated: set status: "completed", proceed to Phase 3.

Resume detection: If status: "in_progress" is found on startup, resume from the first unevaluated skill.

Each skill is evaluated against this checklist:

- [ ] Content overlap with other skills checked
- [ ] Overlap with MEMORY.md / CLAUDE.md checked
- [ ] Freshness of technical references verified (use WebSearch if tool names / CLI flags / APIs are present)
- [ ] Usage frequency considered

Verdict criteria:

VerdictMeaning
KeepUseful and current
ImproveWorth keeping, but specific improvements needed
UpdateReferenced technology is outdated (verify with WebSearch)
RetireLow quality, stale, or cost-asymmetric
Merge into [X]Substantial overlap with another skill; name the merge target

Evaluation is holistic AI judgment — not a numeric rubric. Guiding dimensions:

  • Actionability: code examples, commands, or steps that let you act immediately
  • Scope fit: name, trigger, and content are aligned; not too broad or narrow
  • Uniqueness: value not replaceable by MEMORY.md / CLAUDE.md / another skill
  • Currency: technical references work in the current environment

Reason quality requirements — the reason field must be self-contained and decision-enabling:

  • Do NOT write "unchanged" alone — always restate the core evidence
  • For Retire: state (1) what specific defect was found, (2) what covers the same need instead
    • Bad: "Superseded"
    • Good: "disable-model-invocation: true already set; superseded by continuous-learning-v2 which covers all the same patterns plus confidence scoring. No unique content remains."
  • For Merge: name the target and describe what content to integrate
    • Bad: "Overlaps with X"
    • Good: "42-line thin content; Step 4 of chatlog-to-article already covers the same workflow. Integrate the 'article angle' tip as a note in that skill."
  • For Improve: describe the specific change needed (what section, what action, target size if relevant)
    • Bad: "Too long"
    • Good: "276 lines; Section 'Framework Comparison' (L80–140) duplicates ai-era-architecture-principles; delete it to reach ~150 lines."
  • For Keep (mtime-only change in Quick Scan): restate the original verdict rationale, do not write "unchanged"
    • Bad: "Unchanged"
    • Good: "mtime updated but content unchanged. Unique Python reference explicitly imported by rules/python/; no overlap found."

Phase 3 — Summary Table

Skill7d useVerdictReason

Phase 4 — Consolidation

  1. Retire / Merge: present detailed justification per file before confirming with user:
    • What specific problem was found (overlap, staleness, broken references, etc.)
    • What alternative covers the same functionality (for Retire: which existing skill/rule; for Merge: the target file and what content to integrate)
    • Impact of removal (any dependent skills, MEMORY.md references, or workflows affected)
  2. Improve: present specific improvement suggestions with rationale:
    • What to change and why (e.g., "trim 430→200 lines because sections X/Y duplicate python-patterns")
    • User decides whether to act
  3. Update: present updated content with sources checked
  4. Check MEMORY.md line count; propose compression if >100 lines

Results File Schema

~/.claude/skills/skill-stocktake/results.json:

evaluated_at: Must be set to the actual UTC time of evaluation completion. Obtain via Bash: date -u +%Y-%m-%dT%H:%M:%SZ. Never use a date-only approximation like T00:00:00Z.

json
1{ 2 "evaluated_at": "2026-02-21T10:00:00Z", 3 "mode": "full", 4 "batch_progress": { 5 "total": 80, 6 "evaluated": 80, 7 "status": "completed" 8 }, 9 "skills": { 10 "skill-name": { 11 "path": "~/.claude/skills/skill-name/SKILL.md", 12 "verdict": "Keep", 13 "reason": "Concrete, actionable, unique value for X workflow", 14 "mtime": "2026-01-15T08:30:00Z" 15 } 16 } 17}

Notes

  • Evaluation is blind: the same checklist applies to all skills regardless of origin (ECC, self-authored, auto-extracted)
  • Archive / delete operations always require explicit user confirmation
  • No verdict branching by skill origin

Related Skills

Looking for an alternative to skill-stocktake or building a Categories.official AI Agent? Explore these related open-source MCP Servers.

View All

flags

Logo of facebook
facebook

flags is a feature flag management system that enables developers to check flag states, compare channels, and debug feature behavior differences across release channels.

243.6k
0
Design

extract-errors

Logo of facebook
facebook

extract-errors is a skill that assists in extracting and managing error codes in React applications using yarn extract-errors command.

243.6k
0
Design

fix

Logo of facebook
facebook

fix is a technical skill that resolves lint errors, formatting issues, and ensures code quality in declarative, frontend, and UI projects

243.6k
0
Design

flow

Logo of facebook
facebook

Flow is a type checking system for JavaScript, used to validate React code and ensure consistency across applications

243.6k
0
Design