/interact endpoint — Scrape a page, then call /interact to take actions on it — click buttons, fill forms, navigate deeper, or extract dynamic content. Describe what you want in natural language via prompt, or write Playwright code (Node.js, Python) and Bash (agent-browser) for full control. Sessions persist across calls, with live view and interactive live view URLs for real-time browser streaming. Persistent profiles let you save and reuse browser state (cookies, localStorage) across scrapes. Available in JS, Python, Java, and Rust SDKs.query format — Added query format to the /scrape endpoint — pass a natural-language prompt and get a direct answer back in data.answer.audio format — Added audio format option to scrape responses, returning audio output as a field on the document.onlyCleanContent parameter — Added onlyCleanContent parameter to the /scrape endpoint, which strips navigation, ads, cookie banners, and other non-semantic content from markdown output.fast, auto, ocr) and a maxPages option to control extraction depth and OCR behavior..doc file support — Added support for parsing legacy .doc files.contentType in scrape responses — Added contentType to scrape responses for PDFs and documents.timeout, max_retries, and backoff_factor — these were previously accepted but silently ignored.o3-mini model on extract jobs.time_taken in /v1/map always returning ~0.failed status with an error message and partial data when a crawl-level failure occurs.maxPages not being passed to the PDF extractor — previously, full PDF content was returned while only charging for the limited page count.maxCredits threshold.colors.secondary not being populated.removeBase64Images running after deriveDiff in the transformer pipeline, causing diff issues.ZodError in /v1/search controller.handlebars, path-to-regexp, fast-xml-parser, rollup (CVE-2026-27606), undici, and others.GET /v2/team/activity endpoint for listing recent scrape, crawl, and extract jobs with cursor-based pagination (last 24 hours, up to 100 results per page, filterable by endpoint type).regexOnFullURL parameter on crawl requests to apply includePaths/excludePaths filtering against the full URL including query parameters. Available in JS, Python, Java, and Elixir SDKs.deduplicateSimilarURLs parameter on crawl requests. Available in JS, Python, Java, and Elixir SDKs.extract endpoint — use the /agent endpoint instead. Existing extract methods in JS and Python SDKs are marked deprecated.persistentSession to profile on browser/interact requests (writeMode is now saveChanges). The old parameter name remains functional but is no longer documented.Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.8.0...v2.9.0
Firecrawl v2.9.0 includes browser interaction via /interact, new scrape formats, smarter PDF handling, two new SDKs, and reliability fixes.
Key Features:
/interact — Scrape a page, then call /interact to click buttons, fill forms, navigate, or extract dynamic content using natural language or Playwright/Bash code. Sessions persist across calls with live view URLs and reusable browser profiles./scrape and get direct answers in data.answer.onlyCleanContent Parameter — Strip navigation, ads, and non-semantic content from markdown output.fast, auto, or ocr parsing with maxPages option for fine-grained extraction control.Introduce the new /interact endpoint that turns any scrape into a live browser session where agents can click, type, and navigate using natural language.
Key Features:
Full support for core endpoints including scrape, search, and crawl. Works with Maven, Gradle, and Java 17+.
Key Features:
New PDF parsing engine delivers 3x faster parsing and significantly improved reliability. Rebuilt in Rust, it automatically adapts to any PDF from clean text files to scanned reports and complex layouts.
Key Features:
fast — text-only parsing for maximum performance.auto — new default; starts in fast mode and automatically falls back to OCR when needed, intelligently detecting edge cases like embedded images, graphs, multi-column layouts, and unusual text encodings.ocr — forces OCR parsing for fully image-based or scanned documents.Browser Sandbox gives agents a secure, fully managed browser environment for interactive web automation with no local setup, Chromium installs, or driver compatibility issues. Each session runs in an isolated, disposable sandbox that scales without infrastructure management.
Key Features:
Significantly improved logo extraction accuracy for Branding Format v2, the endpoint for extracting brand identities from websites.
Key Features:
Firecrawl v2.8.0 brings major improvements to agent workflows, developer tooling, and self-hosted deployments across the API and SDKs, including our new Skill.
/agent queries simultaneously, powered by our new Spark 1 Fast model.And much more, check it out below!
Parallel Agents
Execute thousands of /agent queries in parallel with automatic failure handling and intelligent waterfall execution. Powered by Spark 1-Fast for instant retrieval, automatically upgrading to Spark 1 Mini for complex queries requiring full research.
Firecrawl CLI
New command-line interface for Firecrawl with full support for scrape, search, crawl, and map commands. Install with npm install -g firecrawl-cli.
Firecrawl Skill
Enables agents like Claude Cursor, Codex, and OpenCode to use Firecrawl for web scraping and data extraction, installable via npx skills add firecrawl/cli.
Spark Model Family
Three new models powering /agent: Spark 1 Fast for instant retrieval (currently available in Playground), Spark 1 Mini (default) for everyday extraction tasks at 60% lower cost, and Spark 1 Pro for complex multi-domain research requiring maximum accuracy. Spark 1 Pro achieves ~50% recall while Mini delivers ~40% recall, both significantly outperforming tools costing 4-7x more per task.
Firecrawl MCP Server Agent Tools
New firecrawl_agent and firecrawl_agent_status tools for autonomous web data gathering via MCP-enabled agents.
Agent Webhooks
Agent endpoint now supports webhooks for real-time notifications on job completion and progress.
Agent Model Selection
Agent endpoint now accepts a model parameter and includes model info in status responses.
Multi-Arch Docker Images
Self-hosted deployments now support linux/arm64 architecture in addition to amd64.
Sitemap-Only Crawl Mode
New crawl option to exclusively use sitemap URLs without following links.
ignoreCache Map Parameter
New option to bypass cached results when mapping URLs.
Custom Headers for /map
Map endpoint now supports custom request headers.
Background Image Extraction
Scraper now extracts background images from CSS styles.
Improved Error Messages
All user-facing error messages now include detailed explanations to help diagnose issues.
400 for unsupported actions with clear errors when requested actions aren't supported by available engines.og:title or twitter:title when missing.gid parameter when rewriting Google Sheets URLs.robots.txt fetching and parsing.Watcher and WatcherOptions now exported from the SDK entrypoint.jobId for debugging.max_pages handling in crawl requests.lopdf metadata loading performance.html-to-markdown module with multiple bug fixes.firecrawl --api-url http://localhost:3002 for local instances.Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.7.0...v2.8.0
Firecrawl v2.8.0 brings major improvements to agent workflows, developer tooling, and self-hosted deployments across the API and SDKs.
Key Features:
/agent queries simultaneously with automatic failure handling and intelligent waterfall execution. Powered by Spark 1 Fast for instant retrieval, automatically upgrading to Spark 1 Mini for complex queries.Bringing parallel processing to /agent, letting you batch hundreds or thousands of queries simultaneously. What took hours of sequential queries now completes in minutes with automatic failure handling and parallel execution.
Key Features:
/agent queries simultaneously to enrich companies, research competitors, or build datasets at scale.Introducing the Firecrawl Skill and CLI, a new way for AI agents to reliably access real-time web data. With a single install, agents like Claude Code, Antigravity, and OpenCode can access Firecrawl endpoints including scrape, search, crawl, and map.
Key Features:
And a lot more enhacements, check it out below!
Improved Branding Extract
Better logo and color detection for more accurate brand extraction results.
NOQ Scrape System (Experimental)
New scrape pipeline with improved stability and integrated concurrency checks.
Enhanced Redirect Handling
URLs now resolve before mapping, with safer redirect-chain detection and new abort timeouts.
Enterprise Search Parameters
New enterprise-level options available for the /search endpoint.
Integration-Based User Creation
Users can now be automatically created when coming from referring integrations.
minAge Scrape Parameter
Allows requiring a minimum cached age before re-scraping.
Extract Billing Credits
Extract jobs now use the same credit billing system as other endpoints.
Self-Host: Configurable Crawl Concurrency
Self-hosted deployments can now set custom concurrency limits.
Sentry Enhancements
Added Vercel AI integration, configurable sampling rates, and improved exception filtering.
UUIDv7 IDs
All new resources use lexicographically sortable UUIDv7.
maxAge fixes, recursive sitemap support, Vue/Angular router normalization, and skipping subdomain logic for IP addresses./v2/batch/scrape/:jobId/errors endpointdocument event handling.ignoreQueryParameter.Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.6.0...v2.7.0
Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.6.0...v2.7.0
Major release with enterprise features and platform improvements.
Key Features:
python-sdk with model selection by @Chadha93 in https://github.com/firecrawl/firecrawl/pull/2266Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.5.0...v2.6.0
Major release with unified billing, enhanced features, and significant reliability improvements.
Key Features:
We now have the highest quality and most comprehensive web data API available powered by our new semantic index and custom browser stack.
See the benchmarks below:
<img width="1200" height="675" alt="image" src="https://github.com/user-attachments/assets/96a2ba36-0c7f-4fa3-829e-d6ac91b53705" />.xlsx (Excel) files./search pricing updatetracing instead of print.Full diff: https://github.com/firecrawl/firecrawl/compare/v2.4.0...v2.5.0
tracing instead of print by @codetheweb in https://github.com/firecrawl/firecrawl/pull/2324Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.4.0...v2.5.0
Major release delivering the highest quality and most comprehensive web data API with two major infrastructure improvements: a new Semantic Index and a completely custom browser stack.
Key Features:
/v2/x402) — Added a next-gen search API with improved accuracy and speed (#2218)crawl_status_2 RPC (#2239)"cancelled" job status handling and poll interval fixes (#2240, #2265)getDoneJobsOrderedUntil for more stable Redis retrieval (#2258)$ref schema validation edge cases (#2238)docker-compose.yaml issues (#2242, #2252)🔗 Full Changelog: v2.3.0 → v2.4.0
poll_interval param in watcher by @Chadha93 in https://github.com/firecrawl/firecrawl/pull/2155$ref for recursive schema validation by @Chadha93 in https://github.com/firecrawl/firecrawl/pull/2238queue_scrape for nuq schema by @Chadha93 in https://github.com/firecrawl/firecrawl/pull/2272Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.3.0...v2.4.0
Major release featuring open-source scrape-evals benchmark testing 13 web scraping engines on 1,000 URLs, improved full-page extraction with enhanced browser stack, semantic index for faster retrieval of fresh or previously indexed data, 5x cheaper search with auto-recharge credit packs, smarter concurrency and crawl architecture for improved throughput and reliability, and Excel (.xlsx) scraping support for spreadsheets and CSV files.
pkgvuln issueFull Changelog: https://github.com/firecrawl/firecrawl/compare/v2.2.0...v2.3.0