Turn any url into clean data
markdown
)summary
)html
)rawHtml
) (with no modifications)screenshot
, with options like fullPage
, quality
, viewport
)links
)json
) - structured outputprompt
to the endpoint. The llm chooses the structure of the data.
json
format, pass an object inside formats
with the following parameters:
schema
: JSON Schema for the structured output.prompt
: Optional prompt to help guide extraction when a schema is present or when you prefer light guidance.wait
action before/after executing other actions to give enough time for the page to load.
location
object in your request body with the following properties:
country
: ISO 3166-1 alpha-2 country code (e.g., ‘US’, ‘AU’, ‘DE’, ‘JP’). Defaults to ‘US’.languages
: An array of preferred languages and locales for the request in order of priority. Defaults to the language of the specified location.maxAge = 172800000
ms (2 days). If a cached page is newer than this, it’s returned instantly; otherwise, the page is scraped and then cached.maxAge
to 0
.storeInCache
to false
if you don’t want Firecrawl to cache/store results for this request./crawl
endpoint works. It submits a batch scrape job and returns a job ID to check the status of the batch scrape.
The sdk provides 2 methods, synchronous and asynchronous. The synchronous method will return the results of the batch scrape job, while the asynchronous method will return a job ID that you can use to check the status of the batch scrape.
/batch/scrape/{id}
endpoint. This endpoint is meant to be used while the job is still running or right after it has completed as batch scrape jobs expire after 24 hours.