docs/tools/firecrawl.md
OpenClaw can use Firecrawl in three ways:
web_search providerfirecrawl_search and firecrawl_scrapeweb_fetchIt is a hosted extraction/search service that supports bot circumvention and caching, which helps with JS-heavy sites or pages that block plain HTTP fetches.
FIRECRAWL_API_KEY in the gateway environment.{
tools: {
web: {
search: {
provider: "firecrawl",
},
},
},
plugins: {
entries: {
firecrawl: {
enabled: true,
config: {
webSearch: {
apiKey: "FIRECRAWL_API_KEY_HERE",
baseUrl: "https://api.firecrawl.dev",
},
},
},
},
},
}
Notes:
openclaw configure --section web enables the bundled Firecrawl plugin automatically.web_search with Firecrawl supports query and count.sources, categories, or result scraping, use firecrawl_search.baseUrl defaults to hosted Firecrawl at https://api.firecrawl.dev. Self-hosted overrides are allowed only for private/internal endpoints; HTTP is accepted only for those private targets.FIRECRAWL_BASE_URL is the shared env fallback for Firecrawl search and scrape base URLs.{
plugins: {
entries: {
firecrawl: {
enabled: true,
config: {
webFetch: {
apiKey: "FIRECRAWL_API_KEY_HERE",
baseUrl: "https://api.firecrawl.dev",
onlyMainContent: true,
maxAgeMs: 172800000,
timeoutSeconds: 60,
},
},
},
},
},
}
Notes:
plugins.entries.firecrawl.config.webFetch.apiKey or FIRECRAWL_API_KEY).maxAgeMs controls how old cached results can be (ms). Default is 2 days.tools.web.fetch.firecrawl.* config is auto-migrated by openclaw doctor --fix.https://api.firecrawl.dev; self-hosted overrides must resolve to private/internal endpoints.firecrawl_scrape rejects obvious private, loopback, metadata, and non-HTTP(S) target URLs before forwarding them to Firecrawl, matching the web_fetch target-safety contract for explicit Firecrawl scrape calls.firecrawl_scrape reuses the same plugins.entries.firecrawl.config.webFetch.* settings and env vars.
Set plugins.entries.firecrawl.config.webSearch.baseUrl,
plugins.entries.firecrawl.config.webFetch.baseUrl, or FIRECRAWL_BASE_URL
when you run Firecrawl yourself. OpenClaw accepts http:// only for loopback,
private-network, .local, .internal, or .localhost targets. Public custom
hosts are rejected so Firecrawl API keys are not sent to arbitrary endpoints by
accident.
firecrawl_searchUse this when you want Firecrawl-specific search controls instead of generic web_search.
Core parameters:
querycountsourcescategoriesscrapeResultstimeoutSecondsfirecrawl_scrapeUse this for JS-heavy or bot-protected pages where plain web_fetch is weak.
Core parameters:
urlextractModemaxCharsonlyMainContentmaxAgeMsproxystoreInCachetimeoutSecondsFirecrawl exposes a proxy mode parameter for bot circumvention (basic, stealth, or auto).
OpenClaw always uses proxy: "auto" plus storeInCache: true for Firecrawl requests.
If proxy is omitted, Firecrawl defaults to auto. auto retries with stealth proxies if a basic attempt fails, which may use more credits
than basic-only scraping.
web_fetch uses Firecrawlweb_fetch extraction order:
The selection knob is tools.web.fetch.provider. If you omit it, OpenClaw
auto-detects the first ready web-fetch provider from available credentials.
Today the bundled provider is Firecrawl.