Skip to main content When you should use this server
Scrape and extract content from web pages or entire domains.
Perform batch scraping jobs with built-in rate limiting.
Search the web and optionally extract structured content from results.
Conduct deep research using a combination of crawling, search, and LLM-powered summarization.
Generate standardized llms.txt and llms-full.txt files to guide LLM interactions with your site.
Integrate content discovery, crawling, and extraction into AI-driven workflows.
Key features
Web scraping with custom wait times and content filters
Batch scraping with parallel execution
Deep web crawling with configurable depth and limits
Structured data extraction from web pages
LLM-powered research and summarization
Generation of standardized LLM guidance files
Authentication
Method: API Key (FIRECRAWL_API_KEY)
Notes: Store securely in environment variables; available from the Firecrawl dashboard
Endpoints
Remote hosted URL: https://mcp.firecrawl.dev/{FIRECRAWL_API_KEY}/v2/sse
Local via npx: env FIRECRAWL_API_KEY=fc-YOUR_API_KEY npx -y firecrawl-mcp
Manual installation: npm install -g firecrawl-mcp
Setup & usage
Environment variables
FIRECRAWL_API_KEY=fc-YOUR_API_KEY # Required
FIRECRAWL_API_URL=https://custom-endpoint.example.com # Optional, for self-hosted deployments
FIRECRAWL_RETRY_MAX_ATTEMPTS=5 # Optional, default retry configuration
FIRECRAWL_RETRY_INITIAL_DELAY=1000 # Optional, in milliseconds
FIRECRAWL_RETRY_MAX_DELAY=30000 # Optional, in milliseconds
FIRECRAWL_RETRY_BACKOFF_FACTOR=2 # Optional, exponential backoff multiplier
FIRECRAWL_CREDIT_WARNING_THRESHOLD=10 # Optional, percentage threshold for credit warnings
FIRECRAWL_CREDIT_CRITICAL_THRESHOLD=5 # Optional, percentage threshold for critical credit alerts
firecrawl_scrape
Scrape content from a single URL with options for format, wait time, and content filtering.
firecrawl_batch_scrape
Scrape multiple URLs in parallel; returns a batch operation ID for status tracking.
firecrawl_check_batch_status
Check the progress or results of a batch scrape operation.
firecrawl_search
Run a web search; optionally scrape and extract structured content from search results.
firecrawl_crawl
Launch an asynchronous crawl with configurable depth, limits, and external link rules.
Extract structured information from pages using an LLM and a defined schema (supports both cloud and self-hosted LLMs).
firecrawl_deep_research
Perform deep research on a query using crawling, search, and LLM analysis with constraints (depth, time, max URLs).
firecrawl_generate_llmstxt
Generate standardized llms.txt and llms-full.txt files to define how LLMs should interact with a site.
Notes
Supports both cloud-hosted and self-hosted deployments.
Includes robust logging for operation status, rate limits, and credit usage.
Automatic retries and backoff are built-in for error handling.
Returned data respects configured API quotas and credit limits.