Firecrawl allows you to turn entire websites into LLM-ready markdown
Firecrawl is an API service that takes a URL, crawls it, and converts it into clean markdown. We crawl all accessible subpages and give you clean markdown for each. No sitemap required.
We provide an easy to use API with our hosted version. You can find the playground and documentation here. You can also self host the backend if you’d like.
Check out the following resources to get started:
Self-host: To self-host refer to guide here.
To use the API, you need to sign up on Firecrawl and get an API key.
You can find all of Firecrawl’s capabilities and how to use them in our documentation
To scrape a single URL, use the scrape_url
method. It takes the URL as a parameter and returns the scraped data as a dictionary.
SDKs will return the data object directly. cURL will return the payload exactly as shown below.
Used to crawl a URL and all accessible subpages. This submits a crawl job and returns a job ID to check the status of the crawl.
If you’re using cURL or async crawl
functions on SDKs, this will return an ID
where you can use to check the status of the crawl.
Used to check the status of a crawl job and get its result.
The response will be different depending on the status of the crawl. For not completed or large responses exceeding 10MB, a next
URL parameter is provided. You must request this URL to retrieve the next 10MB of data. If the next
parameter is absent, it indicates the end of the crawl data.
With LLM extraction, you can easily extract structured data from any URL. We support pydantic schemas to make it easier for you too. Here is how you to use it:
v1 is only supported on node, python and cURL at this time.
Output:
You can now extract without a schema by just passing a prompt
to the endpoint. The llm chooses the structure of the data.
Output:
Firecrawl allows you to perform various actions on a web page before scraping its content. This is particularly useful for interacting with dynamic content, navigating through pages, or accessing content that requires user interaction.
Here is an example of how to use actions to navigate to google.com, search for Firecrawl, click on the first result, and take a screenshot.
It is important to almost always use the wait
action before/after executing other actions to give enough time for the page to load.
Firecrawl is open source available under the AGPL-3.0 license.
To deliver the best possible product, we offer a hosted version of Firecrawl alongside our open-source offering. The cloud solution allows us to continuously innovate and maintain a high-quality, sustainable service for all users.
Firecrawl Cloud is available at firecrawl.dev and offers a range of features that are not available in the open source version:
We love contributions! Please read our contributing guide before submitting a pull request.
Firecrawl allows you to turn entire websites into LLM-ready markdown
Firecrawl is an API service that takes a URL, crawls it, and converts it into clean markdown. We crawl all accessible subpages and give you clean markdown for each. No sitemap required.
We provide an easy to use API with our hosted version. You can find the playground and documentation here. You can also self host the backend if you’d like.
Check out the following resources to get started:
Self-host: To self-host refer to guide here.
To use the API, you need to sign up on Firecrawl and get an API key.
You can find all of Firecrawl’s capabilities and how to use them in our documentation
To scrape a single URL, use the scrape_url
method. It takes the URL as a parameter and returns the scraped data as a dictionary.
SDKs will return the data object directly. cURL will return the payload exactly as shown below.
Used to crawl a URL and all accessible subpages. This submits a crawl job and returns a job ID to check the status of the crawl.
If you’re using cURL or async crawl
functions on SDKs, this will return an ID
where you can use to check the status of the crawl.
Used to check the status of a crawl job and get its result.
The response will be different depending on the status of the crawl. For not completed or large responses exceeding 10MB, a next
URL parameter is provided. You must request this URL to retrieve the next 10MB of data. If the next
parameter is absent, it indicates the end of the crawl data.
With LLM extraction, you can easily extract structured data from any URL. We support pydantic schemas to make it easier for you too. Here is how you to use it:
v1 is only supported on node, python and cURL at this time.
Output:
You can now extract without a schema by just passing a prompt
to the endpoint. The llm chooses the structure of the data.
Output:
Firecrawl allows you to perform various actions on a web page before scraping its content. This is particularly useful for interacting with dynamic content, navigating through pages, or accessing content that requires user interaction.
Here is an example of how to use actions to navigate to google.com, search for Firecrawl, click on the first result, and take a screenshot.
It is important to almost always use the wait
action before/after executing other actions to give enough time for the page to load.
Firecrawl is open source available under the AGPL-3.0 license.
To deliver the best possible product, we offer a hosted version of Firecrawl alongside our open-source offering. The cloud solution allows us to continuously innovate and maintain a high-quality, sustainable service for all users.
Firecrawl Cloud is available at firecrawl.dev and offers a range of features that are not available in the open source version:
We love contributions! Please read our contributing guide before submitting a pull request.