Global Rank · of 600 Skills
tavily-best-practices AI Agent Skill
View Source: tavily-ai/skills
MediumInstallation
npx skills add tavily-ai/skills --skill tavily-best-practices 7.3K
Installs
Tavily
Tavily is a search API designed for LLMs, enabling AI applications to access real-time web data.
Installation
Python:
pip install tavily-pythonJavaScript:
npm install @tavily/coreSee references/sdk.md for complete SDK reference.
Client Initialization
from tavily import TavilyClient
# Uses TAVILY_API_KEY env var (recommended)
client = TavilyClient()
#With project tracking (for usage organization)
client = TavilyClient(project_id="your-project-id")
# Async client for parallel queries
from tavily import AsyncTavilyClient
async_client = AsyncTavilyClient()Choosing the Right Method
For custom agents/workflows:
| Need | Method |
|---|---|
| Web search results | search() |
| Content from specific URLs | extract() |
| Content from entire site | crawl() |
| URL discovery from site | map() |
For out-of-the-box research:
| Need | Method |
|---|---|
| End-to-end research with AI synthesis | research() |
Quick Reference
search() - Web Search
response = client.search(
query="quantum computing breakthroughs", # Keep under 400 chars
max_results=10,
search_depth="advanced"
)
print(response)Key parameters: query, max_results, search_depth (ultra-fast/fast/basic/advanced), include_domains, exclude_domains, time_range
See references/search.md for complete search reference.
extract() - URL Content Extraction
# Simple one-step extraction
response = client.extract(
urls=["https://docs.example.com"],
extract_depth="advanced"
)
print(response)Key parameters: urls (max 20), extract_depth, query, chunks_per_source (1-5)
See references/extract.md for complete extract reference.
crawl() - Site-Wide Extraction
response = client.crawl(
url="https://docs.example.com",
instructions="Find API documentation pages", # Semantic focus
extract_depth="advanced"
)
print(response)Key parameters: url, max_depth, max_breadth, limit, instructions, chunks_per_source, select_paths, exclude_paths
See references/crawl.md for complete crawl reference.
map() - URL Discovery
response = client.map(
url="https://docs.example.com"
)
print(response)research() - AI-Powered Research
import time
# For comprehensive multi-topic research
result = client.research(
input="Analyze competitive landscape for X in SMB market",
model="pro" # or "mini" for focused queries, "auto" when unsure
)
request_id = result["request_id"]
# Poll until completed
response = client.get_research(request_id)
while response["status"] not in ["completed", "failed"]:
time.sleep(10)
response = client.get_research(request_id)
print(response["content"]) # The research reportKey parameters: input, model ("mini"/"pro"/"auto"), stream, output_schema, citation_format
See references/research.md for complete research reference.
Detailed Guides
For complete parameters, response fields, patterns, and examples:
- references/sdk.md - Python & JavaScript SDK reference, async patterns, Hybrid RAG
- references/search.md - Query optimization, search depth selection, domain filtering, async patterns, post-filtering
- references/extract.md - One-step vs two-step extraction, query/chunks for targeting, advanced mode
- references/crawl.md - Crawl vs Map, instructions for semantic focus, use cases, Map-then-Extract pattern
- references/research.md - Prompting best practices, model selection, streaming, structured output schemas
- references/integrations.md - LangChain, LlamaIndex, CrewAI, Vercel AI SDK, and framework integrations
Installs
Security Audit
View Source
tavily-ai/skills
More from this source
Power your AI Agents with
the best open-source models.
Drop-in OpenAI-compatible API. No data leaves Europe.
Explore Inference APIGLM
GLM 5
$1.00 / $3.20
per M tokens
Kimi
Kimi K2.5
$0.60 / $2.80
per M tokens
MiniMax
MiniMax M2.5
$0.30 / $1.20
per M tokens
Qwen
Qwen3.5 122B
$0.40 / $3.00
per M tokens
How to use this skill
Install tavily-best-practices by running npx skills add tavily-ai/skills --skill tavily-best-practices in your project directory. Run the install command above in your project directory. The skill file will be downloaded from GitHub and placed in your project.
No configuration needed. Your AI agent (Claude Code, Cursor, Windsurf, etc.) automatically detects installed skills and uses them as context when generating code.
The skill enhances your agent's understanding of tavily-best-practices, helping it follow established patterns, avoid common mistakes, and produce production-ready output.
What you get
Skills are plain-text instruction files — not executable code. They encode expert knowledge about frameworks, languages, or tools that your AI agent reads to improve its output. This means zero runtime overhead, no dependency conflicts, and full transparency: you can read and review every instruction before installing.
Compatibility
This skill works with any AI coding agent that supports the skills.sh format, including Claude Code (Anthropic), Cursor, Windsurf, Cline, Aider, and other tools that read project-level context files. Skills are framework-agnostic at the transport level — the content inside determines which language or framework it applies to.
Chat with 100+ AI Models in one App.
Use Claude, ChatGPT, Gemini alongside with EU-Hosted Models like Deepseek, GLM-5, Kimi K2.5 and many more.