Loading source
Pulling the file list, source metadata, and syntax-aware rendering for this listing.
Source from repo
Bulk-extract content from multiple pages of a website via Tavily CLI with depth, breadth, path filters, and markdown output.
Files
Skill
Size
Entrypoint
Format
Open file
Syntax-highlighted preview of this file as included in the skill package.
SKILL.md
1---2name: tavily-crawl3description: |4Crawl websites and extract content from multiple pages via the Tavily CLI. Use this skill when the user wants to crawl a site, download documentation, extract an entire docs section, bulk-extract pages, save a site as local markdown files, or says "crawl", "get all the pages", "download the docs", "extract everything under /docs", "bulk extract", or needs content from many pages on the same domain. Supports depth/breadth control, path filtering, semantic instructions, and saving each page as a local markdown file.5allowed-tools: Bash(tvly *)6---78# tavily crawl910Crawl a website and extract content from multiple pages. Supports saving each page as a local markdown file.1112## Before running any command1314If `tvly` is not found on PATH, install it first:1516```bash17curl -fsSL https://cli.tavily.com/install.sh | bash && tvly login18```1920Do not skip this step or fall back to other tools.2122See [tavily-cli](../tavily-cli/SKILL.md) for alternative install methods and auth options.2324## When to use2526- You need content from many pages on a site (e.g., all `/docs/`)27- You want to download documentation for offline use28- Step 4 in the [workflow](../tavily-cli/SKILL.md): search → extract → map → **crawl** → research2930## Quick start3132```bash33# Basic crawl34tvly crawl "https://docs.example.com" --json3536# Save each page as a markdown file37tvly crawl "https://docs.example.com" --output-dir ./docs/3839# Deeper crawl with limits40tvly crawl "https://docs.example.com" --max-depth 2 --limit 50 --json4142# Filter to specific paths43tvly crawl "https://example.com" --select-paths "/api/.*,/guides/.*" --exclude-paths "/blog/.*" --json4445# Semantic focus (returns relevant chunks, not full pages)46tvly crawl "https://docs.example.com" --instructions "Find authentication docs" --chunks-per-source 3 --json47```4849## Options5051| Option | Description |52|--------|-------------|53| `--max-depth` | Levels deep (1-5, default: 1) |54| `--max-breadth` | Links per page (default: 20) |55| `--limit` | Total pages cap (default: 50) |56| `--instructions` | Natural language guidance for semantic focus |57| `--chunks-per-source` | Chunks per page (1-5, requires `--instructions`) |58| `--extract-depth` | `basic` (default) or `advanced` |59| `--format` | `markdown` (default) or `text` |60| `--select-paths` | Comma-separated regex patterns to include |61| `--exclude-paths` | Comma-separated regex patterns to exclude |62| `--select-domains` | Comma-separated regex for domains to include |63| `--exclude-domains` | Comma-separated regex for domains to exclude |64| `--allow-external / --no-external` | Include external links (default: allow) |65| `--include-images` | Include images |66| `--timeout` | Max wait (10-150 seconds) |67| `-o, --output` | Save JSON output to file |68| `--output-dir` | Save each page as a .md file in directory |69| `--json` | Structured JSON output |7071## Crawl for context vs. data collection7273**For agentic use** (feeding results to an LLM):7475Always use `--instructions` + `--chunks-per-source`. Returns only relevant chunks instead of full pages — prevents context explosion.7677```bash78tvly crawl "https://docs.example.com" --instructions "API authentication" --chunks-per-source 3 --json79```8081**For data collection** (saving to files):8283Use `--output-dir` without `--chunks-per-source` to get full pages as markdown files.8485```bash86tvly crawl "https://docs.example.com" --max-depth 2 --output-dir ./docs/87```8889## Tips9091- **Start conservative** — `--max-depth 1`, `--limit 20` — and scale up.92- **Use `--select-paths`** to focus on the section you need.93- **Use map first** to understand site structure before a full crawl.94- **Always set `--limit`** to prevent runaway crawls.9596## See also9798- [tavily-map](../tavily-map/SKILL.md) — discover URLs before deciding to crawl99- [tavily-extract](../tavily-extract/SKILL.md) — extract individual pages100- [tavily-search](../tavily-search/SKILL.md) — find pages when you don't have a URL101