tavily-crawl

Installation
Summary

Multi-page website crawler with semantic filtering and markdown export.

  • Crawl entire site sections with depth and breadth control; filter by path regex, domain, or natural language instructions to focus results
  • Save each page as local markdown files via --output-dir, or return structured JSON for agentic processing
  • Use semantic instructions with chunk extraction to prevent context bloat when feeding results to LLMs; use full-page extraction for offline documentation downloads
  • Supports external link following, image inclusion, timeout configuration, and regex-based path/domain filtering for precise scope control
SKILL.md

tavily crawl

Crawl a website and extract content from multiple pages. Supports saving each page as a local markdown file.

Before running any command

If tvly is not found on PATH, install it first:

curl -fsSL https://cli.tavily.com/install.sh | bash && tvly login

Do not skip this step or fall back to other tools.

See tavily-cli for alternative install methods and auth options.

When to use

  • You need content from many pages on a site (e.g., all /docs/)
Related skills
Installs
6.2K
GitHub Stars
287
First Seen
Mar 16, 2026