doc-scraper — community doc-scraper, snowflake-dbt-demo, community, ide skills, Claude Code, Cursor, Windsurf

v1.0.0

About this Skill

Perfect for Data Analysis Agents needing streamlined Snowflake documentation management and Markdown conversion. The doc-scraper skill simplifies Snowflake documentation management by scraping docs.snowflake.com sections to Markdown with SQLite caching. It benefits developers working with Snowflake Data Cloud an

sfc-gh-dflippo sfc-gh-dflippo
[0]
[0]
Updated: 3/12/2026

Killer-Skills Review

Decision support comes first. Repository text comes second.

Reviewed Landing Page Review Score: 9/11

Killer-Skills keeps this page indexable because it adds recommendation, limitations, and review signals beyond the upstream repository text.

Original recommendation layer Concrete use-case guidance Explicit limitations and caution Quality floor passed for review Locale and body language aligned
Review Score
9/11
Quality Score
51
Canonical Locale
en
Detected Body Locale
en

Perfect for Data Analysis Agents needing streamlined Snowflake documentation management and Markdown conversion. The doc-scraper skill simplifies Snowflake documentation management by scraping docs.snowflake.com sections to Markdown with SQLite caching. It benefits developers working with Snowflake Data Cloud an

Core Value

Empowers agents to scrape and cache Snowflake documentation from docs.snowflake.com to Markdown using SQLite, facilitating efficient data cloud management and dbt development with protocols like SQL and data formats like Markdown.

Ideal Agent Persona

Perfect for Data Analysis Agents needing streamlined Snowflake documentation management and Markdown conversion.

Capabilities Granted for doc-scraper

Scraping Snowflake documentation for offline access
Caching SQL reference guides for faster lookup
Converting documentation to Markdown for easier integration with dbt projects

! Prerequisites & Limits

  • Requires Python 3 and uv library installation
  • 7-day caching expiration limit
  • Limited to scraping docs.snowflake.com sections

Source Boundary

The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.

After The Review

Decide The Next Action Before You Keep Reading Repository Material

Killer-Skills should not stop at opening repository instructions. It should help you decide whether to install this skill, when to cross-check against trusted collections, and when to move into workflow rollout.

Labs Demo

Browser Sandbox Environment

⚡️ Ready to unleash?

Experience this Agent in a zero-setup browser environment powered by WebContainers. No installation required.

Boot Container Sandbox

FAQ & Installation Steps

These questions and steps mirror the structured data on this page for better search understanding.

? Frequently Asked Questions

What is doc-scraper?

Perfect for Data Analysis Agents needing streamlined Snowflake documentation management and Markdown conversion. The doc-scraper skill simplifies Snowflake documentation management by scraping docs.snowflake.com sections to Markdown with SQLite caching. It benefits developers working with Snowflake Data Cloud an

How do I install doc-scraper?

Run the command: npx killer-skills add sfc-gh-dflippo/snowflake-dbt-demo/doc-scraper. It works with Cursor, Windsurf, VS Code, Claude Code, and 19+ other IDEs.

What are the use cases for doc-scraper?

Key use cases include: Scraping Snowflake documentation for offline access, Caching SQL reference guides for faster lookup, Converting documentation to Markdown for easier integration with dbt projects.

Which IDEs are compatible with doc-scraper?

This skill is compatible with Cursor, Windsurf, VS Code, Trae, Claude Code, OpenClaw, Aider, Codex, OpenCode, Goose, Cline, Roo Code, Kiro, Augment Code, Continue, GitHub Copilot, Sourcegraph Cody, and Amazon Q Developer. Use the Killer-Skills CLI for universal one-command installation.

Are there any limitations for doc-scraper?

Requires Python 3 and uv library installation. 7-day caching expiration limit. Limited to scraping docs.snowflake.com sections.

How To Install

  1. 1. Open your terminal

    Open the terminal or command line in your project directory.

  2. 2. Run the install command

    Run: npx killer-skills add sfc-gh-dflippo/snowflake-dbt-demo/doc-scraper. The CLI will automatically detect your IDE or AI agent and configure the skill.

  3. 3. Start using the skill

    The skill is now active. Your AI agent can use doc-scraper immediately in the current project.

Upstream Repository Material

The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.

Upstream Source

doc-scraper

Install doc-scraper, an AI agent skill for AI agent workflows and automation. Works with Claude Code, Cursor, and Windsurf with one-command setup.

SKILL.md
Readonly
Upstream Repository Material
The section below is imported from the upstream repository and should be treated as secondary evidence. Use the Killer-Skills review above as the primary layer for fit, risk, and installation decisions.
Supporting Evidence

Snowflake Documentation Scraper

Scrapes docs.snowflake.com sections to Markdown with SQLite caching (7-day expiration).

Usage

First time setup (auto-installs uv and doc-scraper):

bash
1python3 .claude/skills/doc-scraper/scripts/doc_scraper.py

Subsequent runs:

bash
1doc-scraper --output-dir=./snowflake-docs 2doc-scraper --output-dir=./snowflake-docs --base-path="/en/sql-reference/" 3doc-scraper --output-dir=./snowflake-docs --spider-depth=2

Command Options

OptionDefaultDescription
--output-dirRequiredOutput directory for scraped docs
--base-path/en/migrations/URL section to scrape
--spider-depth1Link depth: 0=seeds, 1=+links, 2=+2nd
--limitNoneCap URLs (for testing)
--dry-run-Preview without writing

Output

sql
1output-dir/ 2├── SKILL.md # Auto-generated index 3├── scraper_config.yaml # Editable config (auto-created) 4├── .cache/ # SQLite cache (auto-managed) 5└── en/migrations/*.md # Scraped pages with frontmatter

Configuration

Auto-created at {output-dir}/scraper_config.yaml:

yaml
1rate_limiting: 2 max_concurrent_threads: 4 3spider: 4 max_pages: 1000 5 allowed_paths: ["/en/"] 6scraped_pages: 7 expiration_days: 7

Troubleshooting

IssueSolution
Too many pagesLower --spider-depth or edit config
Missing pagesIncrease --spider-depth
Cache corruptionDelete {output-dir}/.cache/ (rare)

Related Skills

Looking for an alternative to doc-scraper or another community skill for your workflow? Explore these related open-source skills.

View All

openclaw-release-maintainer

Logo of openclaw
openclaw

Your own personal AI assistant. Any OS. Any Platform. The lobster way. 🦞

333.8k
0
AI

widget-generator

Logo of f
f

Generate customizable widget plugins for the prompts.chat feed system

149.6k
0
AI

flags

Logo of vercel
vercel

The React Framework

138.4k
0
Browser

pr-review

Logo of pytorch
pytorch

Tensors and Dynamic neural networks in Python with strong GPU acceleration

98.6k
0
Developer