seed — for Claude Code EngramR, community, for Claude Code, ide skills, no-confirm, ralph N, EXECUTE, Target, ARGUMENTS, required

v1.0

このスキルについて

自動での重複検出とアーカイブ フォルダの作成が必要なファイル管理エージェントに最適です。 ローカライズされた概要: Add a source file to the processing queue. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

機能

Target: $ARGUMENTS
File path (required unless listing or --all): the source file to seed
Add a source file to the processing queue
Checks for duplicates, creates archive folder, moves source from inbox, creates extract task, and
Triggers on /seed, /seed [file], queue this for processing

# 主なトピック

achousal achousal
[1]
[0]
更新日: 3/4/2026

Skill Overview

Start with fit, limitations, and setup before diving into the repository.

自動での重複検出とアーカイブ フォルダの作成が必要なファイル管理エージェントに最適です。 ローカライズされた概要: Add a source file to the processing queue. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

このスキルを使用する理由

エージェントにファイル処理タスクの自動化を可能にします。重複ファイルの確認、アーカイブ フォルダの作成、キューの更新など、ファイル パスの解析や --all、--no-confirm などのフラグを使用するプロトコルを使用します。

おすすめ

自動での重複検出とアーカイブ フォルダの作成が必要なファイル管理エージェントに最適です。

実現可能なユースケース for seed

開発者向けのファイル管理タスクの自動化
--all フラグを使用した複数ファイルの重複検出
--no-confirm フラグを使用した対話式の重複確認プロンプトのスキップ

! セキュリティと制限

  • --all フラグを使用しない限り、ファイル パスを引数として必要とする
  • --all フラグを使用する場合、inbox/ ディレクトリ内のファイルのみに限定される
  • デーモン P2.5 に依存して動作する

About The Source

The section below is adapted from the upstream repository. Use it as supporting material alongside the fit, use-case, and installation summary on this page.

Labs デモ

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

FAQ とインストール手順

These questions and steps mirror the structured data on this page for better search understanding.

? よくある質問

seed とは何ですか?

自動での重複検出とアーカイブ フォルダの作成が必要なファイル管理エージェントに最適です。 ローカライズされた概要: Add a source file to the processing queue. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

seed はどうやって導入しますか?

次のコマンドを実行してください: npx killer-skills add achousal/EngramR。Cursor、Windsurf、VS Code、Claude Code など19以上のIDEで使えます。

seed の主な用途は?

主な用途は次のとおりです: 開発者向けのファイル管理タスクの自動化, --all フラグを使用した複数ファイルの重複検出, --no-confirm フラグを使用した対話式の重複確認プロンプトのスキップ。

seed に対応するIDEは?

このスキルは Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer に対応しています。統一された導入には Killer-Skills CLI を使えます。

seed に制限はありますか?

--all フラグを使用しない限り、ファイル パスを引数として必要とする. --all フラグを使用する場合、inbox/ ディレクトリ内のファイルのみに限定される. デーモン P2.5 に依存して動作する.

このスキルの導入方法

  1. 1. ターミナルを開く

    プロジェクトディレクトリでターミナルまたはコマンドラインを開きます。

  2. 2. インストールコマンドを実行

    npx killer-skills add achousal/EngramR を実行してください。CLI がIDEまたはエージェントを自動検出し、スキルを設定します。

  3. 3. スキルを使い始める

    このスキルはすぐに有効になります。現在のプロジェクトで seed をすぐ使えます。

! Source Notes

This page is still useful for installation and source reference. Before using it, compare the fit, limitations, and upstream repository notes above.

Upstream Repository Material

The section below is adapted from the upstream repository. Use it as supporting material alongside the fit, use-case, and installation summary on this page.

Upstream Source

seed

ローカライズされた概要: Add a source file to the processing queue. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

SKILL.md
Readonly
Upstream Repository Material
The section below is adapted from the upstream repository. Use it as supporting material alongside the fit, use-case, and installation summary on this page.
Upstream Source

EXECUTE NOW

Target: $ARGUMENTS

Parse arguments:

  • File path (required unless listing or --all): the source file to seed
  • --all: loop over ALL files in inbox/, seeding each with --no-confirm. After seeding all files, report count and recommend /ralph N.
  • --no-confirm: skip interactive duplicate-confirmation prompt. If a duplicate is detected, auto-skip and log "Skipped {file}: duplicate detected (--no-confirm)". All other seed operations proceed normally. Used by daemon P2.5 and /seed --all.
  • --methods-only: restrict downstream extraction to methodology-comparisons and design-patterns only. Sets scope: "methods_only" on the extract task and queue entry. Use for cross-domain papers that are methodologically relevant but not disease-specific.

The target MUST be a file path (unless --all). If no target and no --all, list inbox/ contents and ask which to seed.

START NOW. Seed the source file into the processing queue.


Step 1: Validate Source

Confirm the target file exists. If it does not, check common locations:

  • inbox/{filename}
  • Subdirectories of inbox/

If the file cannot be found, report error and stop:

ERROR: Source file not found: {path}
Checked: {locations checked}

Read the file to understand:

  • Content type: what kind of material is this? (research article, documentation, transcript, etc.)
  • Size: line count (affects chunking decisions in /reduce)
  • Format: markdown, plain text, structured data

Step 2: Duplicate Detection

Check if this source has already been processed. Two levels of detection:

2a. Filename Match

Search the queue file and archive folders for matching source names:

bash
1SOURCE_NAME=$(basename "$FILE" .md | tr ' ' '-' | tr '[:upper:]' '[:lower:]') 2 3# Check queue for existing entry 4# Search in ops/queue.yaml, ops/queue/queue.yaml, or ops/queue/queue.json 5grep -l "$SOURCE_NAME" ops/queue*.yaml ops/queue/*.yaml ops/queue/*.json 2>/dev/null 6 7# Check archive folders 8ls -d ops/queue/archive/*-${SOURCE_NAME}* 2>/dev/null

2b. Content Similarity (if semantic search available)

If semantic search is available (qmd MCP tools or CLI), check for content overlap:

mcp__qmd__search query="claims from {source filename}" limit=5

Or via keyword search in the notes/ directory:

bash
1grep -rl "{key terms from source title}" notes/ 2>/dev/null | head -5

2c. Report Duplicates

If either check finds a match:

  • Show what was found (filename match or content overlap)
  • If --no-confirm is set: auto-skip this file. Log "Skipped {file}: duplicate detected (--no-confirm)" and stop cleanly. Do NOT proceed with seeding.
  • Otherwise: Ask: "This source may have been processed before. Proceed anyway? (y/n)"
    • If the user declines, stop cleanly
    • If the user confirms (or no duplicate found), continue

Step 3: Create Archive Structure

Create the archive folder. The date-prefixed folder name ensures uniqueness.

bash
1DATE=$(date -u +"%Y-%m-%d") 2SOURCE_BASENAME=$(basename "$FILE" .md | tr ' ' '-' | tr '[:upper:]' '[:lower:]') 3ARCHIVE_DIR="ops/queue/archive/${DATE}-${SOURCE_BASENAME}" 4mkdir -p "$ARCHIVE_DIR"

The archive folder serves two purposes:

  1. Permanent home for the source file (moved from inbox/)
  2. Destination for task files after batch completion (/archive-batch moves them here)

Step 4: Move Source to Archive

Move the source file from its current location to the archive folder. This is the claiming step — once moved, the source is owned by this processing batch.

inbox/ sources get moved:

bash
1if [[ "$FILE" == *"inbox"* ]]; then 2 mv "$FILE" "$ARCHIVE_DIR/" 3 FINAL_SOURCE="$ARCHIVE_DIR/$(basename "$FILE")" 4fi

Sources outside inbox/ stay in place:

bash
1# Living docs (like configuration files) stay where they are 2# Archive folder is still created for task files 3FINAL_SOURCE="$FILE"

Use $FINAL_SOURCE in the task file — this is the path all downstream phases reference.

Why move immediately: All references (task files, claims' Source footers) use the final archived path from the start. No path updates needed later. If it is in inbox/, it is unclaimed. Claimed sources live in archive.

Step 5: Determine Claim Numbering

Find the highest existing claim number across the queue and archive to ensure globally unique claim IDs.

bash
1# Check queue for highest claim number in file references 2QUEUE_MAX=$(grep -oE '[0-9]{3}\.md' ops/queue*.yaml ops/queue/*.yaml 2>/dev/null | \ 3 grep -oE '[0-9]{3}' | sort -n | tail -1) 4QUEUE_MAX=${QUEUE_MAX:-0} 5 6# Check archive for highest claim number 7ARCHIVE_MAX=$(find ops/queue/archive -name "*-[0-9][0-9][0-9].md" 2>/dev/null | \ 8 grep -v summary | sed 's/.*-\([0-9][0-9][0-9]\)\.md/\1/' | sort -n | tail -1) 9ARCHIVE_MAX=${ARCHIVE_MAX:-0} 10 11# Next claim starts after the highest 12NEXT_CLAIM_START=$((QUEUE_MAX > ARCHIVE_MAX ? QUEUE_MAX + 1 : ARCHIVE_MAX + 1))

Claim numbers are globally unique and never reused across batches. This ensures every claim file name ({source}-{NNN}.md) is unique vault-wide.

Step 5b: Content Depth Detection and Silent Auto-Enrichment

Read the source file's frontmatter for content_depth.

Auto-enrichment (silent): If the source has a DOI (source_url contains 10.) AND no content_depth field AND no ## Abstract section with content, attempt silent enrichment:

bash
1set -a && source _code/.env 2>/dev/null && set +a && uv run --directory _code python -c " 2import json, sys; sys.path.insert(0, 'src') 3from engram_r.stub_enricher import enrich_single_doi 4result = enrich_single_doi('{DOI}') 5print(json.dumps(result)) 6"
  • Success (abstract found): Apply enrichment to the source file, set content_depth: abstract, continue with scope: abstract_only.
  • Failure (no abstract): Log one line: "Abstract fetch failed for {DOI}. Extracting from available metadata only." Set content_depth: stub, continue with scope: stub.
  • No interactive prompt. Auto-enrichment is always silent.

After auto-enrichment (or if content_depth already set):

  • content_depth: abstract -- auto-set scope: abstract_only and inform:

    [Content Depth] Source is abstract-only. Extraction scope set to abstract_only
    (claims, evidence, open-questions only -- no methods or design patterns).
    
  • content_depth: stub -- inform and continue:

    [Content Depth] Source is stub (no abstract). Extraction limited to title-level claim only.
    
  • content_depth: full_text or absent with no DOI -- normal processing, scope: full (unless overridden by --methods-only).

Write the detected content_depth into the extract task file frontmatter.

Step 6: Create Extract Task File

Write the task file to ops/queue/${SOURCE_BASENAME}.md:

markdown
1--- 2id: "{SOURCE_BASENAME}" 3type: extract 4source: "{FINAL_SOURCE}" 5original_path: "{original file path before move}" 6archive_folder: "{ARCHIVE_DIR}" 7created: "{UTC timestamp}" 8next_claim_start: {NEXT_CLAIM_START} 9scope: "{full | methods_only | abstract_only — default full}" 10content_depth: "{stub | abstract | full_text — detected in Step 5b}" 11--- 12 13# Extract claims from {source filename} 14 15## Source 16Original: {original file path} 17Archived: {FINAL_SOURCE} 18Size: {line count} lines 19Content type: {detected type} 20 21## Scope 22{If --methods-only: "Methods only -- extract methodology-comparisons and design-patterns only" | Otherwise: "Full document"} 23 24## Acceptance Criteria 25- Extract claims, implementation ideas, tensions, and testable hypotheses 26- Duplicate check against notes/ during extraction 27- Near-duplicates create enrichment tasks (do not skip) 28- Each output type gets appropriate handling 29 30## Execution Notes 31(filled by /reduce) 32 33## Outputs 34(filled by /reduce)

Step 7: Update Queue

Add the extract task entry to the queue file.

For YAML queues (ops/queue.yaml):

yaml
1- id: {SOURCE_BASENAME} 2 type: extract 3 status: pending 4 source: "{FINAL_SOURCE}" 5 file: "{SOURCE_BASENAME}.md" 6 created: "{UTC timestamp}" 7 next_claim_start: {NEXT_CLAIM_START}

For JSON queues (ops/queue/queue.json):

json
1{ 2 "id": "{SOURCE_BASENAME}", 3 "type": "extract", 4 "status": "pending", 5 "source": "{FINAL_SOURCE}", 6 "file": "{SOURCE_BASENAME}.md", 7 "created": "{UTC timestamp}", 8 "next_claim_start": {NEXT_CLAIM_START}, 9 "scope": "{full | methods_only — matches task file scope}" 10}

If no queue file exists: Create one with the appropriate schema header (phase_order definitions) and this first task entry.

Step 8: Report

--=={ seed }==--

Seeded: {SOURCE_BASENAME}
Source: {original path} -> {FINAL_SOURCE}
Archive folder: {ARCHIVE_DIR}
Size: {line count} lines
Content type: {detected type}

Task file: ops/queue/{SOURCE_BASENAME}.md
Claims will start at: {NEXT_CLAIM_START}
Claim files will be: {SOURCE_BASENAME}-{NNN}.md (unique across vault)
Queue: updated with extract task

Next steps:
  /ralph 1 --batch {SOURCE_BASENAME}     (extract claims)

Why This Skill Exists

Manual queue management is error-prone. This skill:

  • Ensures consistent task file format across batches
  • Handles claim numbering automatically (globally unique)
  • Checks for duplicates before creating unnecessary work
  • Moves sources to their permanent archive location immediately
  • Provides clear next steps for the user

Naming Convention

Task files use the source basename for human readability:

  • Task file: {source-basename}.md
  • Claim files: {source-basename}-{NNN}.md
  • Summary: {source-basename}-summary.md
  • Archive folder: {date}-{source-basename}/

Claim numbers (NNN) are globally unique across all batches, ensuring every filename is unique vault-wide. This is required because wiki links resolve by filename, not path.

Source Handling Patterns

inbox/ source (most common):

inbox/research/article.md
    | /seed
    v
ops/queue/archive/2026-01-30-article/article.md  <- source moved here
ops/queue/article.md                               <- task file created

Living doc (outside inbox/):

CLAUDE.md -> stays as CLAUDE.md (no move)
ops/queue/archive/2026-01-30-claude-md/           <- folder still created
ops/queue/claude-md.md                             <- task file created

When /archive-batch runs later, it moves task files into the existing archive folder and generates a summary.


Edge Cases

Source outside inbox/: Works — source stays in place, archive folder is created for task files only.

No queue file: Create ops/queue/queue.yaml (or .json) with schema header and this first entry.

Large source (2500+ lines): Note in output: "Large source ({N} lines) -- /reduce will chunk automatically."

Source is a URL or non-file: Report error: "/seed requires a file path."


Batch Mode (--all)

When --all is set:

  1. List all files in inbox/ (including subdirectories).
  2. For each file, run Steps 1-8 with --no-confirm (skip duplicate confirmation).
  3. Track results: seeded count, skipped count (duplicates), error count.
  4. After all files processed, report:
--=={ seed --all }==--

Seeded: {seeded_count} files
Skipped: {skipped_count} (duplicates)
Errors: {error_count}

Files seeded:
  {list of seeded filenames}

Next: /ralph {seeded_count} to process all extractions
  1. Do NOT auto-run /ralph. The user controls pacing.

Critical Constraints

never:

  • Skip duplicate detection (prevents wasted processing)
  • Move a source that is not in inbox/ (living docs stay in place)
  • Reuse claim numbers from previous batches (globally unique is required)
  • Create a task file without updating the queue (both must happen together)

always:

  • Ask before proceeding when duplicates are detected
  • Create the archive folder even for living docs (task files need it)
  • Use the archived path (not original) in the task file for inbox/ sources
  • Report next steps clearly so the user knows what to do next
  • Compute next_claim_start from both queue AND archive (not just one)

関連スキル

Looking for an alternative to seed or another community skill for your workflow? Explore these related open-source skills.

すべて表示

openclaw-release-maintainer

Logo of openclaw
openclaw

ローカライズされた概要: 🦞 # OpenClaw Release Maintainer Use this skill for release and publish-time workflow. It covers ai, assistant, crustacean workflows. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

333.8k
0
AI

widget-generator

Logo of f
f

ローカライズされた概要: Generate customizable widget plugins for the prompts.chat feed system # Widget Generator Skill This skill guides creation of widget plugins for prompts.chat. It covers ai, artificial-intelligence, awesome-list workflows. This AI agent skill supports Claude Code, Cursor, and Windsurf

149.6k
0
AI

flags

Logo of vercel
vercel

ローカライズされた概要: The React Framework # Feature Flags Use this skill when adding or changing framework feature flags in Next.js internals. It covers blog, browser, compiler workflows. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

138.4k
0
ブラウザ

pr-review

Logo of pytorch
pytorch

ローカライズされた概要: Usage Modes No Argument If the user invokes /pr-review with no arguments, do not perform a review. It covers autograd, deep-learning, gpu workflows. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

98.6k
0
開発者