Firecrawl thoroughly crawls websites, ensuring comprehensive data extraction while handling complex web infrastructure. Here’s how it works:Documentation Index
Fetch the complete documentation index at: https://docs.firecrawl.dev/llms.txt
Use this file to discover all available pages before exploring further.
- URL Analysis: Begins with a specified URL, identifying links by looking at the sitemap and then crawling the website. If no sitemap is found, it will crawl the website following the links.
- Recursive Traversal: Recursively follows each link to uncover all subpages.
- Content Scraping: Gathers content from every visited page while handling any complexities like JavaScript rendering or rate limits.
- Result Compilation: Converts collected data into clean markdown or structured output, perfect for LLM processing or any other task.
Crawling
/crawl endpoint
Used to crawl a URL and all accessible subpages. This submits a crawl job and returns a job ID to check the status of the crawl.Installation
Usage
Job ID Response
If you are not using the sdk or prefer to use webhook or a different polling method, you can set thewait_until_done to false.
This will return a jobId.
For cURL, /crawl will always return a jobId where you can use to check the status of the crawl.

