seed — for Claude Code EngramR, community, for Claude Code, ide skills, no-confirm, ralph N, EXECUTE, Target, ARGUMENTS, required

v1.0

이 스킬 정보

자동으로 중복 파일을 감지하고 보관 폴더를 생성해야 하는 파일 관리 에이전트에게 적합합니다. 현지화된 요약: Add a source file to the processing queue. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

기능

Target: $ARGUMENTS
File path (required unless listing or --all): the source file to seed
Add a source file to the processing queue
Checks for duplicates, creates archive folder, moves source from inbox, creates extract task, and
Triggers on /seed, /seed [file], queue this for processing

# 핵심 주제

achousal achousal
[1]
[0]
업데이트: 3/4/2026

Skill Overview

Start with fit, limitations, and setup before diving into the repository.

자동으로 중복 파일을 감지하고 보관 폴더를 생성해야 하는 파일 관리 에이전트에게 적합합니다. 현지화된 요약: Add a source file to the processing queue. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

이 스킬을 사용하는 이유

에이전트가 파일 처리 작업을 자동화할 수 있도록 합니다. 이는 파일 경로 파싱과 --all 및 --no-confirm와 같은 플래그를 사용하는 프로토콜을 통해 중복 파일 확인, 보관 폴더 생성 및 큐 업데이트 등을 수행할 수 있습니다.

최적의 용도

자동으로 중복 파일을 감지하고 보관 폴더를 생성해야 하는 파일 관리 에이전트에게 적합합니다.

실행 가능한 사용 사례 for seed

개발자를 위한 파일 관리 작업 자동화
--all 플래그를 사용한 여러 파일의 중복 감지
--no-confirm 플래그를 사용한 대화식 중복 확인 프롬프트 생략

! 보안 및 제한 사항

  • --all 플래그를 사용하지 않는 한 파일 경로를 인수로 필요로 함
  • --all 플래그를 사용하는 경우 inbox/ 디렉토리 내의 파일만으로 제한됨
  • 데몬 P2.5에 의존하여 작동함

About The Source

The section below comes from the upstream repository. Use it as supporting material alongside the fit, use-case, and installation summary on this page.

Labs 데모

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

FAQ 및 설치 단계

These questions and steps mirror the structured data on this page for better search understanding.

? 자주 묻는 질문

seed은 무엇인가요?

자동으로 중복 파일을 감지하고 보관 폴더를 생성해야 하는 파일 관리 에이전트에게 적합합니다. 현지화된 요약: Add a source file to the processing queue. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

seed은 어떻게 설치하나요?

다음 명령을 실행하세요: npx killer-skills add achousal/EngramR. Cursor, Windsurf, VS Code, Claude Code와 19개 이상의 다른 IDE에서 동작합니다.

seed은 어디에 쓰이나요?

주요 활용 사례는 다음과 같습니다: 개발자를 위한 파일 관리 작업 자동화, --all 플래그를 사용한 여러 파일의 중복 감지, --no-confirm 플래그를 사용한 대화식 중복 확인 프롬프트 생략.

seed 와 호환되는 IDE는 무엇인가요?

이 스킬은 Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer 와 호환됩니다. 통합 설치에는 Killer-Skills CLI를 사용하세요.

seed에 제한 사항이 있나요?

--all 플래그를 사용하지 않는 한 파일 경로를 인수로 필요로 함. --all 플래그를 사용하는 경우 inbox/ 디렉토리 내의 파일만으로 제한됨. 데몬 P2.5에 의존하여 작동함.

이 스킬 설치 방법

  1. 1. 터미널 열기

    프로젝트 디렉터리에서 터미널 또는 명령줄을 여세요.

  2. 2. 설치 명령 실행

    npx killer-skills add achousal/EngramR 를 실행하세요. CLI가 IDE 또는 에이전트를 자동으로 감지하고 스킬을 설정합니다.

  3. 3. 스킬 사용 시작

    스킬이 이제 활성화되었습니다. 현재 프로젝트에서 seed을 바로 사용할 수 있습니다.

! Source Notes

This page is still useful for installation and source reference. Before using it, compare the fit, limitations, and upstream repository notes above.

Upstream Repository Material

The section below comes from the upstream repository. Use it as supporting material alongside the fit, use-case, and installation summary on this page.

Upstream Source

seed

현지화된 요약: Add a source file to the processing queue. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

SKILL.md
Readonly
Upstream Repository Material
The section below comes from the upstream repository. Use it as supporting material alongside the fit, use-case, and installation summary on this page.
Upstream Source

EXECUTE NOW

Target: $ARGUMENTS

Parse arguments:

  • File path (required unless listing or --all): the source file to seed
  • --all: loop over ALL files in inbox/, seeding each with --no-confirm. After seeding all files, report count and recommend /ralph N.
  • --no-confirm: skip interactive duplicate-confirmation prompt. If a duplicate is detected, auto-skip and log "Skipped {file}: duplicate detected (--no-confirm)". All other seed operations proceed normally. Used by daemon P2.5 and /seed --all.
  • --methods-only: restrict downstream extraction to methodology-comparisons and design-patterns only. Sets scope: "methods_only" on the extract task and queue entry. Use for cross-domain papers that are methodologically relevant but not disease-specific.

The target MUST be a file path (unless --all). If no target and no --all, list inbox/ contents and ask which to seed.

START NOW. Seed the source file into the processing queue.


Step 1: Validate Source

Confirm the target file exists. If it does not, check common locations:

  • inbox/{filename}
  • Subdirectories of inbox/

If the file cannot be found, report error and stop:

ERROR: Source file not found: {path}
Checked: {locations checked}

Read the file to understand:

  • Content type: what kind of material is this? (research article, documentation, transcript, etc.)
  • Size: line count (affects chunking decisions in /reduce)
  • Format: markdown, plain text, structured data

Step 2: Duplicate Detection

Check if this source has already been processed. Two levels of detection:

2a. Filename Match

Search the queue file and archive folders for matching source names:

bash
1SOURCE_NAME=$(basename "$FILE" .md | tr ' ' '-' | tr '[:upper:]' '[:lower:]') 2 3# Check queue for existing entry 4# Search in ops/queue.yaml, ops/queue/queue.yaml, or ops/queue/queue.json 5grep -l "$SOURCE_NAME" ops/queue*.yaml ops/queue/*.yaml ops/queue/*.json 2>/dev/null 6 7# Check archive folders 8ls -d ops/queue/archive/*-${SOURCE_NAME}* 2>/dev/null

2b. Content Similarity (if semantic search available)

If semantic search is available (qmd MCP tools or CLI), check for content overlap:

mcp__qmd__search query="claims from {source filename}" limit=5

Or via keyword search in the notes/ directory:

bash
1grep -rl "{key terms from source title}" notes/ 2>/dev/null | head -5

2c. Report Duplicates

If either check finds a match:

  • Show what was found (filename match or content overlap)
  • If --no-confirm is set: auto-skip this file. Log "Skipped {file}: duplicate detected (--no-confirm)" and stop cleanly. Do NOT proceed with seeding.
  • Otherwise: Ask: "This source may have been processed before. Proceed anyway? (y/n)"
    • If the user declines, stop cleanly
    • If the user confirms (or no duplicate found), continue

Step 3: Create Archive Structure

Create the archive folder. The date-prefixed folder name ensures uniqueness.

bash
1DATE=$(date -u +"%Y-%m-%d") 2SOURCE_BASENAME=$(basename "$FILE" .md | tr ' ' '-' | tr '[:upper:]' '[:lower:]') 3ARCHIVE_DIR="ops/queue/archive/${DATE}-${SOURCE_BASENAME}" 4mkdir -p "$ARCHIVE_DIR"

The archive folder serves two purposes:

  1. Permanent home for the source file (moved from inbox/)
  2. Destination for task files after batch completion (/archive-batch moves them here)

Step 4: Move Source to Archive

Move the source file from its current location to the archive folder. This is the claiming step — once moved, the source is owned by this processing batch.

inbox/ sources get moved:

bash
1if [[ "$FILE" == *"inbox"* ]]; then 2 mv "$FILE" "$ARCHIVE_DIR/" 3 FINAL_SOURCE="$ARCHIVE_DIR/$(basename "$FILE")" 4fi

Sources outside inbox/ stay in place:

bash
1# Living docs (like configuration files) stay where they are 2# Archive folder is still created for task files 3FINAL_SOURCE="$FILE"

Use $FINAL_SOURCE in the task file — this is the path all downstream phases reference.

Why move immediately: All references (task files, claims' Source footers) use the final archived path from the start. No path updates needed later. If it is in inbox/, it is unclaimed. Claimed sources live in archive.

Step 5: Determine Claim Numbering

Find the highest existing claim number across the queue and archive to ensure globally unique claim IDs.

bash
1# Check queue for highest claim number in file references 2QUEUE_MAX=$(grep -oE '[0-9]{3}\.md' ops/queue*.yaml ops/queue/*.yaml 2>/dev/null | \ 3 grep -oE '[0-9]{3}' | sort -n | tail -1) 4QUEUE_MAX=${QUEUE_MAX:-0} 5 6# Check archive for highest claim number 7ARCHIVE_MAX=$(find ops/queue/archive -name "*-[0-9][0-9][0-9].md" 2>/dev/null | \ 8 grep -v summary | sed 's/.*-\([0-9][0-9][0-9]\)\.md/\1/' | sort -n | tail -1) 9ARCHIVE_MAX=${ARCHIVE_MAX:-0} 10 11# Next claim starts after the highest 12NEXT_CLAIM_START=$((QUEUE_MAX > ARCHIVE_MAX ? QUEUE_MAX + 1 : ARCHIVE_MAX + 1))

Claim numbers are globally unique and never reused across batches. This ensures every claim file name ({source}-{NNN}.md) is unique vault-wide.

Step 5b: Content Depth Detection and Silent Auto-Enrichment

Read the source file's frontmatter for content_depth.

Auto-enrichment (silent): If the source has a DOI (source_url contains 10.) AND no content_depth field AND no ## Abstract section with content, attempt silent enrichment:

bash
1set -a && source _code/.env 2>/dev/null && set +a && uv run --directory _code python -c " 2import json, sys; sys.path.insert(0, 'src') 3from engram_r.stub_enricher import enrich_single_doi 4result = enrich_single_doi('{DOI}') 5print(json.dumps(result)) 6"
  • Success (abstract found): Apply enrichment to the source file, set content_depth: abstract, continue with scope: abstract_only.
  • Failure (no abstract): Log one line: "Abstract fetch failed for {DOI}. Extracting from available metadata only." Set content_depth: stub, continue with scope: stub.
  • No interactive prompt. Auto-enrichment is always silent.

After auto-enrichment (or if content_depth already set):

  • content_depth: abstract -- auto-set scope: abstract_only and inform:

    [Content Depth] Source is abstract-only. Extraction scope set to abstract_only
    (claims, evidence, open-questions only -- no methods or design patterns).
    
  • content_depth: stub -- inform and continue:

    [Content Depth] Source is stub (no abstract). Extraction limited to title-level claim only.
    
  • content_depth: full_text or absent with no DOI -- normal processing, scope: full (unless overridden by --methods-only).

Write the detected content_depth into the extract task file frontmatter.

Step 6: Create Extract Task File

Write the task file to ops/queue/${SOURCE_BASENAME}.md:

markdown
1--- 2id: "{SOURCE_BASENAME}" 3type: extract 4source: "{FINAL_SOURCE}" 5original_path: "{original file path before move}" 6archive_folder: "{ARCHIVE_DIR}" 7created: "{UTC timestamp}" 8next_claim_start: {NEXT_CLAIM_START} 9scope: "{full | methods_only | abstract_only — default full}" 10content_depth: "{stub | abstract | full_text — detected in Step 5b}" 11--- 12 13# Extract claims from {source filename} 14 15## Source 16Original: {original file path} 17Archived: {FINAL_SOURCE} 18Size: {line count} lines 19Content type: {detected type} 20 21## Scope 22{If --methods-only: "Methods only -- extract methodology-comparisons and design-patterns only" | Otherwise: "Full document"} 23 24## Acceptance Criteria 25- Extract claims, implementation ideas, tensions, and testable hypotheses 26- Duplicate check against notes/ during extraction 27- Near-duplicates create enrichment tasks (do not skip) 28- Each output type gets appropriate handling 29 30## Execution Notes 31(filled by /reduce) 32 33## Outputs 34(filled by /reduce)

Step 7: Update Queue

Add the extract task entry to the queue file.

For YAML queues (ops/queue.yaml):

yaml
1- id: {SOURCE_BASENAME} 2 type: extract 3 status: pending 4 source: "{FINAL_SOURCE}" 5 file: "{SOURCE_BASENAME}.md" 6 created: "{UTC timestamp}" 7 next_claim_start: {NEXT_CLAIM_START}

For JSON queues (ops/queue/queue.json):

json
1{ 2 "id": "{SOURCE_BASENAME}", 3 "type": "extract", 4 "status": "pending", 5 "source": "{FINAL_SOURCE}", 6 "file": "{SOURCE_BASENAME}.md", 7 "created": "{UTC timestamp}", 8 "next_claim_start": {NEXT_CLAIM_START}, 9 "scope": "{full | methods_only — matches task file scope}" 10}

If no queue file exists: Create one with the appropriate schema header (phase_order definitions) and this first task entry.

Step 8: Report

--=={ seed }==--

Seeded: {SOURCE_BASENAME}
Source: {original path} -> {FINAL_SOURCE}
Archive folder: {ARCHIVE_DIR}
Size: {line count} lines
Content type: {detected type}

Task file: ops/queue/{SOURCE_BASENAME}.md
Claims will start at: {NEXT_CLAIM_START}
Claim files will be: {SOURCE_BASENAME}-{NNN}.md (unique across vault)
Queue: updated with extract task

Next steps:
  /ralph 1 --batch {SOURCE_BASENAME}     (extract claims)

Why This Skill Exists

Manual queue management is error-prone. This skill:

  • Ensures consistent task file format across batches
  • Handles claim numbering automatically (globally unique)
  • Checks for duplicates before creating unnecessary work
  • Moves sources to their permanent archive location immediately
  • Provides clear next steps for the user

Naming Convention

Task files use the source basename for human readability:

  • Task file: {source-basename}.md
  • Claim files: {source-basename}-{NNN}.md
  • Summary: {source-basename}-summary.md
  • Archive folder: {date}-{source-basename}/

Claim numbers (NNN) are globally unique across all batches, ensuring every filename is unique vault-wide. This is required because wiki links resolve by filename, not path.

Source Handling Patterns

inbox/ source (most common):

inbox/research/article.md
    | /seed
    v
ops/queue/archive/2026-01-30-article/article.md  <- source moved here
ops/queue/article.md                               <- task file created

Living doc (outside inbox/):

CLAUDE.md -> stays as CLAUDE.md (no move)
ops/queue/archive/2026-01-30-claude-md/           <- folder still created
ops/queue/claude-md.md                             <- task file created

When /archive-batch runs later, it moves task files into the existing archive folder and generates a summary.


Edge Cases

Source outside inbox/: Works — source stays in place, archive folder is created for task files only.

No queue file: Create ops/queue/queue.yaml (or .json) with schema header and this first entry.

Large source (2500+ lines): Note in output: "Large source ({N} lines) -- /reduce will chunk automatically."

Source is a URL or non-file: Report error: "/seed requires a file path."


Batch Mode (--all)

When --all is set:

  1. List all files in inbox/ (including subdirectories).
  2. For each file, run Steps 1-8 with --no-confirm (skip duplicate confirmation).
  3. Track results: seeded count, skipped count (duplicates), error count.
  4. After all files processed, report:
--=={ seed --all }==--

Seeded: {seeded_count} files
Skipped: {skipped_count} (duplicates)
Errors: {error_count}

Files seeded:
  {list of seeded filenames}

Next: /ralph {seeded_count} to process all extractions
  1. Do NOT auto-run /ralph. The user controls pacing.

Critical Constraints

never:

  • Skip duplicate detection (prevents wasted processing)
  • Move a source that is not in inbox/ (living docs stay in place)
  • Reuse claim numbers from previous batches (globally unique is required)
  • Create a task file without updating the queue (both must happen together)

always:

  • Ask before proceeding when duplicates are detected
  • Create the archive folder even for living docs (task files need it)
  • Use the archived path (not original) in the task file for inbox/ sources
  • Report next steps clearly so the user knows what to do next
  • Compute next_claim_start from both queue AND archive (not just one)

관련 스킬

Looking for an alternative to seed or another community skill for your workflow? Explore these related open-source skills.

모두 보기

openclaw-release-maintainer

Logo of openclaw
openclaw

현지화된 요약: 🦞 # OpenClaw Release Maintainer Use this skill for release and publish-time workflow. It covers ai, assistant, crustacean workflows. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

333.8k
0
인공지능

widget-generator

Logo of f
f

현지화된 요약: Generate customizable widget plugins for the prompts.chat feed system # Widget Generator Skill This skill guides creation of widget plugins for prompts.chat . It covers ai, artificial-intelligence, awesome-list workflows. This AI agent skill supports Claude Code, Cursor, and Windsurf

149.6k
0
인공지능

flags

Logo of vercel
vercel

현지화된 요약: The React Framework # Feature Flags Use this skill when adding or changing framework feature flags in Next.js internals. It covers blog, browser, compiler workflows. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

138.4k
0
브라우저

pr-review

Logo of pytorch
pytorch

현지화된 요약: Usage Modes No Argument If the user invokes /pr-review with no arguments, do not perform a review . It covers autograd, deep-learning, gpu workflows. This AI agent skill supports Claude Code, Cursor, and Windsurf workflows.

98.6k
0
개발자