releases.shpreview
Firecrawl/GitHub Releases

GitHub Releases

$npx -y @buildinternet/releases show firecrawl-github-releases
Mon
Wed
Fri
AprMayJunJulAugSepOctNovDecJanFebMarApr
Less
More
Releases2Avg0/wkVersionsv2.8.0 → v2.9.0
Apr 10, 2026

Firecrawl v2.9.0

Improvements

  • Browser Interaction via /interact endpoint — Scrape a page, then call /interact to take actions on it — click buttons, fill forms, navigate deeper, or extract dynamic content. Describe what you want in natural language via prompt, or write Playwright code (Node.js, Python) and Bash (agent-browser) for full control. Sessions persist across calls, with live view and interactive live view URLs for real-time browser streaming. Persistent profiles let you save and reuse browser state (cookies, localStorage) across scrapes. Available in JS, Python, Java, and Rust SDKs.
  • query format — Added query format to the /scrape endpoint — pass a natural-language prompt and get a direct answer back in data.answer.
  • audio format — Added audio format option to scrape responses, returning audio output as a field on the document.
  • onlyCleanContent parameter — Added onlyCleanContent parameter to the /scrape endpoint, which strips navigation, ads, cookie banners, and other non-semantic content from markdown output.
  • PDF parsing modes — Added PDF parsing modes (fast, auto, ocr) and a maxPages option to control extraction depth and OCR behavior.
  • Java and Elixir SDKs — Added official Java and Elixir SDKs with full v2 API support.
  • Legacy .doc file support — Added support for parsing legacy .doc files.
  • Wikimedia engine — Added a dedicated engine for scraping Wikipedia and Wikimedia pages with improved output quality.
  • contentType in scrape responses — Added contentType to scrape responses for PDFs and documents.
  • PDF pipeline improvements — Improved PDF pipeline with better table detection, header/footer stripping, mixed PDF handling, inline image parsing, and magic byte detection.
  • Branding extraction — Improved branding extraction to skip hidden DOM elements for cleaner output.
  • HTML-to-markdown performance — Improved HTML-to-markdown conversion performance and fixed code blocks losing content during conversion.
  • Concurrency queue — New concurrency queue system with reconciler and backfill for more reliable job scheduling.
  • Rust SDK v2 — Added v2 API namespace with agent support to the Rust SDK.
  • Fixed Python SDK parameters timeout, max_retries, and backoff_factor — these were previously accepted but silently ignored.
  • Capped job timeouts at 48 hours to prevent runaway jobs from consuming resources.
  • Added retry limits to prevent scrape loops.
  • Binary content types are now rejected early in the scrape pipeline to avoid wasted processing.

Fixes

  • Fixed empty responses when using the o3-mini model on extract jobs.
  • Fixed revoked API keys remaining valid for up to 10 minutes after deletion.
  • Fixed a race condition in extract jobs that caused "Job not found" crashes.
  • Fixed time_taken in /v1/map always returning ~0.
  • Fixed crawl status responses now surfacing a failed status with an error message and partial data when a crawl-level failure occurs.
  • Fixed maxPages not being passed to the PDF extractor — previously, full PDF content was returned while only charging for the limited page count.
  • Fixed free request credits being incorrectly consumed and billed on agent jobs exceeding the maxCredits threshold.
  • Fixed dashboard displaying incorrect concurrency limits due to stale reads.
  • Fixed branding colors.secondary not being populated.
  • Fixed removeBase64Images running after deriveDiff in the transformer pipeline, causing diff issues.
  • Fixed GCS fetch using wrong row index for cache info lookups.
  • Fixed unhandled ZodError in /v1/search controller.
  • Resolved multiple CVEs across dependencies including handlebars, path-to-regexp, fast-xml-parser, rollup (CVE-2026-27606), undici, and others.
  • Hardened the Playwright service against SSRF attacks.

API

  • Added GET /v2/team/activity endpoint for listing recent scrape, crawl, and extract jobs with cursor-based pagination (last 24 hours, up to 100 results per page, filterable by endpoint type).
  • Added regexOnFullURL parameter on crawl requests to apply includePaths/excludePaths filtering against the full URL including query parameters. Available in JS, Python, Java, and Elixir SDKs.
  • Added deduplicateSimilarURLs parameter on crawl requests. Available in JS, Python, Java, and Elixir SDKs.
  • Deprecated the extract endpoint — use the /agent endpoint instead. Existing extract methods in JS and Python SDKs are marked deprecated.
  • Renamed persistentSession to profile on browser/interact requests (writeMode is now saveChanges). The old parameter name remains functional but is no longer documented.

New Contributors

Contributors

  • @nickscamara
  • @mogery
  • @amplitudesxd
  • @abimaelmartell
  • @ericciarla
  • @rafaelsideguide
  • @delong3
  • @devhims
  • @Chadha93
  • @tomsideguide
  • @charlietlamb
  • @developersdigest
  • @micahstairs
  • @rhys-firecrawl
  • @firecrawl-spring
  • @devin-ai-integration
  • @misza-one
  • @madmikeross
  • @rowinsg
  • @Bortlesboat
  • @dagecko
  • @cokemine
  • @paulonasc

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.8.0...v2.9.0

Feb 3, 2026

Firecrawl v2.8.0 is here!

Firecrawl v2.8.0 brings major improvements to agent workflows, developer tooling, and self-hosted deployments across the API and SDKs, including our new Skill.

  • Parallel Agents for running thousands of /agent queries simultaneously, powered by our new Spark 1 Fast model.
  • Firecrawl CLI with full support for scrape, search, crawl, and map commands.
  • Firecrawl Skill for enabling AI agents (Claude Code, Codex, OpenCode) to use Firecrawl autonomously.
  • Three new models powering /agent: Spark 1 Fast for instant retrieval (currently only available in Playground), Spark 1 Mini for complex research queries, and Spark 1 Pro for advanced extraction tasks.
  • Agent enhancements including webhooks, model selection, and new MCP Server tools.
  • Platform-wide performance improvements including faster search execution and optimized Redis calls.
  • SDK improvements including Zod v4 compatibility.

And much more, check it out below!

New Features

  • Parallel Agents
    Execute thousands of /agent queries in parallel with automatic failure handling and intelligent waterfall execution. Powered by Spark 1-Fast for instant retrieval, automatically upgrading to Spark 1 Mini for complex queries requiring full research.

  • Firecrawl CLI
    New command-line interface for Firecrawl with full support for scrape, search, crawl, and map commands. Install with npm install -g firecrawl-cli.

  • Firecrawl Skill
    Enables agents like Claude Cursor, Codex, and OpenCode to use Firecrawl for web scraping and data extraction, installable via npx skills add firecrawl/cli.

  • Spark Model Family
    Three new models powering /agent: Spark 1 Fast for instant retrieval (currently available in Playground), Spark 1 Mini (default) for everyday extraction tasks at 60% lower cost, and Spark 1 Pro for complex multi-domain research requiring maximum accuracy. Spark 1 Pro achieves ~50% recall while Mini delivers ~40% recall, both significantly outperforming tools costing 4-7x more per task.

  • Firecrawl MCP Server Agent Tools
    New firecrawl_agent and firecrawl_agent_status tools for autonomous web data gathering via MCP-enabled agents.

  • Agent Webhooks
    Agent endpoint now supports webhooks for real-time notifications on job completion and progress.

  • Agent Model Selection
    Agent endpoint now accepts a model parameter and includes model info in status responses.

  • Multi-Arch Docker Images
    Self-hosted deployments now support linux/arm64 architecture in addition to amd64.

  • Sitemap-Only Crawl Mode
    New crawl option to exclusively use sitemap URLs without following links.

  • ignoreCache Map Parameter
    New option to bypass cached results when mapping URLs.

  • Custom Headers for /map
    Map endpoint now supports custom request headers.

  • Background Image Extraction
    Scraper now extracts background images from CSS styles.

  • Improved Error Messages
    All user-facing error messages now include detailed explanations to help diagnose issues.


API Improvements

  • Search without concurrency limits — scrapes in search now execute directly without queue overhead.
  • Return 400 for unsupported actions with clear errors when requested actions aren't supported by available engines.
  • Job ID now included in search metadata for easier tracking.
  • Metadata responses now include detected timezone.
  • Backfill metadata title from og:title or twitter:title when missing.
  • Preserve gid parameter when rewriting Google Sheets URLs.
  • Fixed v2 path in batch scrape status pagination.
  • Validate team ownership when appending to existing crawls.
  • Screenshots with custom viewport or quality settings now bypass cache.
  • Optimized Redis calls across endpoints.
  • Reduced excessive robots.txt fetching and parsing.
  • Minimum request timeout parameter now configurable.

SDK Improvements

JavaScript SDK

  • Zod v4 Compatibility — schema conversion now works with Zod v4 with improved error detection.
  • Watcher ExportsWatcher and WatcherOptions now exported from the SDK entrypoint.
  • Agent Webhook Support — new webhook options for agent calls.
  • Error Retry Polling — SDK retries polling after transient errors.
  • Job ID in Exceptions — error exceptions now include jobId for debugging.

Python SDK

  • Manual pagination helpers for iterating through results.
  • Agent webhook support added to agent client.
  • Agent endpoint now accepts model selection parameter.
  • Metadata now includes concurrency limit information.
  • Fixed max_pages handling in crawl requests.

Dashboard Improvements

  • Dark mode is now supported.
  • On the usage page, you can now view credit usage broken down by day.
  • On the activity logs page, you can now filter by the API key that was used.
  • The "images" output format is now supported in the Playground.
  • All admins can now manage their team's subscriptions.

Quality & Performance

  • Skip markdown conversion checks for large HTML documents.
  • Export Google Docs as HTML instead of PDF for improved performance.
  • Improved branding format with better logo detection and error messages for PDFs and documents.
  • Improved lopdf metadata loading performance.
  • Updated html-to-markdown module with multiple bug fixes.
  • Increased markdown service body limit and added request ID logging.
  • Better Sentry filtering for cancelled jobs and engine errors.
  • Fixed extract race conditions and RabbitMQ poison pill handling.
  • Centralized Firecrawl configuration across the codebase.
  • Multiple security vulnerability fixes, including CVE-2025-59466 and lodash prototype pollution.

Self-Hosted Improvements

  • CLI custom API URL support via firecrawl --api-url http://localhost:3002 for local instances.
  • ARM64 Docker support via multi-arch images for Apple Silicon and ARM servers.
  • Fixed docker-compose database credentials out of the box.
  • Fixed Playwright service startup caused by Chromium path issues.
  • Updated Node.js to major version 22 instead of a pinned minor.
  • Added RabbitMQ health check endpoint.
  • Fixed PostgreSQL port exposure in docker-compose.

New Contributors

  • @gemyago
  • @loganaden
  • @pcgeek86
  • @dmlarionov

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.7.0...v2.8.0

What's Changed

Dec 5, 2025

Firecrawl v2.7.0 is here!

  • ZDR Search support for enterprise customers.
  • Improved Branding Format with better detection.
  • Partner Integrations API now in closed beta.
  • Faster and more accurate screenshots.
  • Self-hosted improvements

And a lot more enhacements, check it out below!

New Features

  • Improved Branding Extract
    Better logo and color detection for more accurate brand extraction results.

  • NOQ Scrape System (Experimental)
    New scrape pipeline with improved stability and integrated concurrency checks.

  • Enhanced Redirect Handling
    URLs now resolve before mapping, with safer redirect-chain detection and new abort timeouts.

  • Enterprise Search Parameters
    New enterprise-level options available for the /search endpoint.

  • Integration-Based User Creation
    Users can now be automatically created when coming from referring integrations.

  • minAge Scrape Parameter
    Allows requiring a minimum cached age before re-scraping.

  • Extract Billing Credits
    Extract jobs now use the same credit billing system as other endpoints.

  • Self-Host: Configurable Crawl Concurrency
    Self-hosted deployments can now set custom concurrency limits.

  • Sentry Enhancements
    Added Vercel AI integration, configurable sampling rates, and improved exception filtering.

  • UUIDv7 IDs
    All new resources use lexicographically sortable UUIDv7.

API Improvements

  • DNS Resolution Errors Now Return 200 for more consistent failure handling.
  • Improved URL Mapping Logic including sitemap maxAge fixes, recursive sitemap support, Vue/Angular router normalization, and skipping subdomain logic for IP addresses.
  • Partial Results for Multi-Source Search instead of failing all sources.
  • Concurrency Metadata Added to scrape job responses.
  • Enhanced Metrics including total wait time, LLM usage, and format details.
  • Batch Scrape Upgrades
    • Added missing /v2/batch/scrape/:jobId/errors endpoint
    • Fixed pagination off-by-one bug
  • More Robust Error Handling for PDF/document engines, pydantic parsing, Zod validation, URL validation, and billing edge cases.

SDK Improvements

JavaScript SDK

  • Returns job ID from synchronous methods.
  • Improved WebSocket document event handling.
  • Fixed types, Deno WS, and added support for ignoreQueryParameter.
  • Version bump with internal cleanup.

Python SDK

  • Added extra metadata fields.
  • Improved batch validation handling.

Quality & Performance

  • Reduced log file size and improved tmp file cleanup.
  • Updated Express version and patched vulnerable packages.
  • Disabled markdown conversion for sitemap scrapes for improved performance.
  • Better precrawl logging and formatting.
  • Skip URL rewriting for published Google Docs.
  • Prevent empty cookie headers during webhook callbacks.

Self-Hosted Improvements

  • Disabled concurrency limit enforcement for self-hosted mode.
  • PostgreSQL credentials now configurable via environment variables.
  • Docker-compose build instructions fixed.

👥 New Contributors

  • @omahs
  • @davidkhala
  • @DraPraks
  • @devhims

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.6.0...v2.7.0

What's Changed

New Contributors

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.6.0...v2.7.0

Nov 14, 2025

Highlights

  • Unified Billing Model - Credits and tokens merged into single system. Extract now uses credits (15 tokens = 1 credit), existing tokens work everywhere.
  • Full Release of Branding Format - Full support across Playground, MCP, JS and Python SDKs.
  • Change Tracking - Faster and more reliable detection of web page content updates.
  • Reliability and Speed Improvements - All endpoints significantly faster with improved reliability.
  • Instant Credit Purchases - Buy credit packs directly from dashboard without waiting for auto-recharge.
  • Improved Markdown Parsing - Enhanced markdown conversion and main content extraction accuracy.
  • Core Stability Fixes - Fixed change-tracking issues, PDF timeouts, and improved error handling.

What's Changed

New Contributors

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.5.0...v2.6.0

Oct 30, 2025
v2.5.0 - The World's Best Web Data API

We now have the highest quality and most comprehensive web data API available powered by our new semantic index and custom browser stack.

See the benchmarks below:

<img width="1200" height="675" alt="image" src="https://github.com/user-attachments/assets/96a2ba36-0c7f-4fa3-829e-d6ac91b53705" />

New Features

  • Implemented scraping for .xlsx (Excel) files.
  • Introduced new crawl architecture and NUQ concurrency tracking system.
  • Per-owner/group concurrency limiting + dynamic concurrency calculation.
  • Added group backlog handling and improved group operations.
  • Added /search pricing update
  • Added team flag to skip country check.
  • Always populate NUQ metrics for improved observability.
  • New test-site app for improved CI testing.
  • Extract metadata from document head for richer output.

Enhancements & Improvements

  • Improved blocklist loading and unsupported site error messages.
  • Updated x402-express version.
  • Improved includePaths handling for subdomains.
  • Updated self-hosted search to use DuckDuckGo.
  • JS & Python SDKs no longer require API key for self-hosted deployments.
  • Python SDK timeout handling improvements.
  • Rust client now uses tracing instead of print.
  • Reduced noise in auto-recharge Slack notifications.

Fixes

  • Ensured crawl robots.txt warnings surface reliably.
  • Resolved concurrency deadlocks and duplicate job handling.
  • Fixed search country defaults and pricing logic bugs.
  • Fixed port conflicts in harness environments.
  • Fixed viewport dimension support and screenshot behavior in Playwright.
  • Resolved CI test flakiness (playwright cache, prod tests).

👋 New Contributors

  • @delong3
  • @c4nc
  • @codetheweb

Full diff: https://github.com/firecrawl/firecrawl/compare/v2.4.0...v2.5.0

What's Changed

New Contributors

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.4.0...v2.5.0

Oct 13, 2025

New Features

  • New PDF Search Category - You can now search for only pdfs via our v2/search endpoints by specifying .pdf category
  • Gemini 2.5 Flash CLI Image Editor — Create and edit images directly in the CLI using Firecrawl + Gemini 2.5 Flash integration (#2172)
  • x402 Search Endpoint (/v2/x402) — Added a next-gen search API with improved accuracy and speed (#2218)
  • RabbitMQ Event System — Firecrawl jobs now support event-based communication and prefetching from Postgres (#2230, #2233)
  • Improved Crawl Status API — More accurate and real-time crawl status reporting using the new crawl_status_2 RPC (#2239)
  • Low-Results & Robots.txt Warnings — Users now receive clear feedback when crawls are limited by robots.txt or yield few results (#2248)
  • Enhanced Tracing (OpenTelemetry) — Much-improved distributed tracing for better observability across services (#2219)
  • Metrics & Analytics — Added request-level metrics for both Scrape and Search endpoints (#2216)
  • Self-Hosted Webhook Support — Webhooks can now be delivered to private IP addresses for self-hosted environments (#2232)

Improvements

  • Reduced Docker Image Size — Playwright service image size reduced by 1 GB by only installing Chromium (#2210)
  • Python SDK Enhancements — Added "cancelled" job status handling and poll interval fixes (#2240, #2265)
  • Faster Node SDK Timeouts — Axios timeouts now propagate correctly, improving reliability under heavy loads (#2235)
  • Improved Crawl Parameter Previews — Enhanced prompts and validation for crawl parameter previews (#2220)
  • Zod Schema Validation — Stricter API parameter validation with rejection of extra fields (#2058)
  • Better Redis Job Handling — Fixed edge cases in getDoneJobsOrderedUntil for more stable Redis retrieval (#2258)
  • Markdown & YouTube Fixes — Fixed YouTube cache and empty markdown summary bugs (#2226, #2261)
  • Updated Docs & Metadata — README updates and new metadata fields added to the JS SDK (#2250, #2254)
  • Improved API Port Configuration — The API now respects environment-defined ports (#2209)

Fixes

  • Fixed recursive $ref schema validation edge cases (#2238)
  • Fixed enum arrays being incorrectly converted to objects (#2224)
  • Fixed harness timeouts and self-hosted docker-compose.yaml issues (#2242, #2252)

New Contributors

🔗 Full Changelog: v2.3.0 → v2.4.0

What's Changed

New Contributors

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.3.0...v2.4.0

Sep 19, 2025

New Features

  • YouTube Support: You can now get YouTube transcripts
  • Enterprise Auto-Recharge: Added enterprise support for auto-recharge
  • odt and .rtf: Now support odt and rtf file parsing
  • Docx Parsing: 50x faster docx parsing
  • K8s Deployment: Added NuQ worker deployment example
  • Self Host: Tons of improvements for our self host users

Improvements & Fixes

  • Stability: Fixed timeout race condition, infinite scrape loop, and location query bug
  • Tooling: Replaced ts-prune with knip, updated pnpm with minimumReleaseAge
  • Docs: Added Rust to CONTRIBUTING and fixed typos
  • Security: Fixed pkgvuln issue

What's Changed

New Contributors

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.2.0...v2.3.0

Sep 12, 2025

Features

  • MCP version 3 is live. Stable support for cloud mcp with HTTP Transport and SSE modes. Compatible with v2 and v1 from.
  • Webhooks: Now we support signatures + extract support + event failures
  • Map is now 15x faster + supports more urls
  • Search reliability improvements
  • Usage is now tracked by API Key
  • Support for additional locations (CA, CZ, IL, IN, IT, PL, and PT)
  • Queue status endpoint
  • Added maxPages parameter to v2 scrape API for pdf parsing

Improvements

  • API:
    • New /team/queue-status endpoint.
    • Added nuq feature.
    • Added VIASOCKET integration.
    • Historical credit/token usage endpoints with expanded data.
  • Student Program: Support for more universities + students to get free credits through our student program
  • Map: 15x faster and increased the limit to 100k
  • Scrape API: Added maxPages parameter for PDF parser.
  • Python SDK:
    • Added get_queue_status to aio + normalization of docs in search results.
    • SDKs: Added next cursor pagination and integration param support.
  • Infrastructure: Added static IP proxy pool + proxy location support.
  • Webhooks: Implemented signatures, refactored sending, added scrape error events.
  • Performance: Optimized map, converted Rust natives to single NAPI library.
  • CI/CD: Revamped CI, added pre-commit hooks, cross-platform harness.

🐛 Fixes

  • Corrected concurrency limit scaling.
  • Fixed search result links/descriptions and retry mechanism for empty results.
  • Re-signed expired screenshot URLs.
  • Trimmed null chars from PDF titles + fixed encoding.
  • Fixed sitemap parsing and added .gz sitemap support.
  • Fixed js-sdk zod-to-json-schema import.
  • Fixed webhook data format regression.
  • Improved credit handling in account object.

🛠️ Chores & Other

  • Removed unused dependencies, updated CONTRIBUTING.md.
  • Added debug logging, ignored scripts during CI build.
  • Various dependency bumps and build improvements.

🔗 Full Changelog: v2.1.0...v2.2.0

What's Changed

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.1.0...v2.2.0

Aug 29, 2025

Firecrawl v2.1.0 is here!

✨ New Features

  • Search Categories: Filter search results by specific categories using the categories parameter:
    • github: Search within GitHub repositories, code, issues, and documentation
    • research: Search academic and research websites (arXiv, Nature, IEEE, PubMed, etc.)
    • More coming soon
  • Image Extraction: Added image extraction support to the v2 scrape endpoint.
  • Data Attribute Scraping: Now supports extraction of data-* attributes.
  • Hash-Based Routing: Crawl endpoints now handle hash-based routes.
  • Improved Google Drive Scraping: Added ability to scrape TXT, PDF, and Sheets from Google Drive.
  • PDF Enhancements: Extracts PDF titles and shows them in metadata.
  • API Enhancements:
    • Map endpoint supports up to 100k results.
  • Helm Chart: Initial Helm chart added for Firecrawl deployment.
  • Security: Improved protection against XFF spoofing.

🛠 Fixes

  • Fixed UTF-8 encoding in Google search scraper.
  • Restored crawl status in preview mode.
  • Fixed missing methods in Python SDK.
  • Corrected JSON response handling for v2 search with scrapeOptions.formats.
  • Fixed field population for credits_billed in v0 scrape.
  • Improved document field overlay in v2 search.

👥 New Contributors

  • @kelter-antunes
  • @vishkrish200
  • @ieedan

🔗 Full Changelog

What's Changed

New Contributors

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v2.0.1...v2.1.0

Aug 26, 2025

This release fixes the "SSRF Vulnerability via malicious webhook" security advisory. It is recommended that people using the self-hosted version of Firecrawl update to v2.0.1 immediately. More info in the advisory: https://github.com/firecrawl/firecrawl/security/advisories/GHSA-p2wg-prhf-jx79

Aug 19, 2025

Introducing v2.0.0

Key Improvements

  • Faster by default: Requests are cached with maxAge defaulting to 2 days, and sensible defaults like blockAds, skipTlsVerification, and removeBase64Images are enabled.

  • New summary format: You can now specify "summary" as a format to directly receive a concise summary of the page content.

  • Updated JSON extraction: JSON extraction and change tracking now use an object format: { type: "json", prompt, schema }. The old "extract" format has been renamed to "json".

  • Enhanced screenshot options: Use the object form: { type: "screenshot", fullPage, quality, viewport }.

  • New search sources: Search across "news" and "images" in addition to web results by setting the sources parameter.

  • Smart crawling with prompts: Pass a natural-language prompt to crawl and the system derives paths/limits automatically. Use the new crawl-params-preview endpoint to inspect the derived options before starting a job.

Quick migration checklist

  • Replace v1 client usage with v2 clients:
    • JS: const firecrawl = new Firecrawl({ apiKey: 'fc-YOUR-API-KEY' })
    • Python: firecrawl = Firecrawl(api_key='fc-YOUR-API-KEY')
    • API: use the new https://api.firecrawl.dev/v2/ endpoints.
  • Update formats:
    • Use "summary" where needed
    • JSON mode: Use { type: "json", prompt, schema } for JSON extraction
    • Screenshot and Screenshot@fullPage: Use screenshot object format when specifying options
  • Adopt standardized async flows in the SDKs:
    • Crawls: startCrawl + getCrawlStatus (or crawl waiter)
    • Batch: startBatchScrape + getBatchScrapeStatus (or batchScrape waiter)
    • Extract: startExtract + getExtractStatus (or extract waiter)
  • Crawl options mapping (see below)
  • Check crawl prompt with crawl-params-preview

SDK surface (v2)

JS/TS

Method name changes (v1 → v2)

Scrape, Search, and Map

v1 (FirecrawlApp)v2 (Firecrawl)
scrapeUrl(url, ...)scrape(url, options?)
search(query, ...)search(query, options?)
mapUrl(url, ...)map(url, options?)

Crawling

v1v2
crawlUrl(url, ...)crawl(url, options?) (waiter)
asyncCrawlUrl(url, ...)startCrawl(url, options?)
checkCrawlStatus(id, ...)getCrawlStatus(id)
cancelCrawl(id)cancelCrawl(id)
checkCrawlErrors(id)getCrawlErrors(id)

Batch Scraping

v1v2
batchScrapeUrls(urls, ...)batchScrape(urls, opts?) (waiter)
asyncBatchScrapeUrls(urls, ...)startBatchScrape(urls, opts?)
checkBatchScrapeStatus(id, ...)getBatchScrapeStatus(id)
checkBatchScrapeErrors(id)getBatchScrapeErrors(id)

Extraction

v1v2
extract(urls?, params?)extract(args)
asyncExtract(urls, params?)startExtract(args)
getExtractStatus(id)getExtractStatus(id)

Other / Removed

v1v2
generateLLMsText(...)(not in v2 SDK)
checkGenerateLLMsTextStatus(id)(not in v2 SDK)
crawlUrlAndWatch(...)watcher(jobId, ...)
batchScrapeUrlsAndWatch(...)watcher(jobId, ...)

Type name changes (v1 → v2)

Core Document Types

v1v2
FirecrawlDocumentDocument
FirecrawlDocumentMetadataDocumentMetadata

Scrape, Search, and Map Types

v1v2
ScrapeParamsScrapeOptions
ScrapeResponseDocument
SearchParamsSearchRequest
SearchResponseSearchData
MapParamsMapOptions
MapResponseMapData

Crawl Types

v1v2
CrawlParamsCrawlOptions
CrawlStatusResponseCrawlJob

Batch Operations

v1v2
BatchScrapeStatusResponseBatchScrapeJob

Action Types

v1v2
ActionActionOption

Error Types

v1v2
FirecrawlErrorSdkError
ErrorResponseErrorDetails

Python (sync)

Method name changes (v1 → v2)

Scrape, Search, and Map

v1v2
scrape_url(...)scrape(...)
search(...)search(...)
map_url(...)map(...)

Crawling

v1v2
crawl_url(...)crawl(...) (waiter)
async_crawl_url(...)start_crawl(...)
check_crawl_status(...)get_crawl_status(...)
cancel_crawl(...)cancel_crawl(...)

Batch Scraping

v1v2
batch_scrape_urls(...)batch_scrape(...) (waiter)
async_batch_scrape_urls(...)start_batch_scrape(...)
get_batch_scrape_status(...)get_batch_scrape_status(...)
get_batch_scrape_errors(...)get_batch_scrape_errors(...)

Extraction

v1v2
extract(...)extract(...)
start_extract(...)start_extract(...)
get_extract_status(...)get_extract_status(...)

Other / Removed

v1v2
generate_llms_text(...)(not in v2 SDK)
get_generate_llms_text_status(...)(not in v2 SDK)
watch_crawl(...)watcher(job_id, ...)

Python (async)

  • AsyncFirecrawl mirrors the same methods (all awaitable).

Formats and scrape options

  • Use string formats for basics: "markdown", "html", "rawHtml", "links", "summary".
  • Instead of parsePDF use parsers: [ { "type": "pdf" } | "pdf" ].
  • Use object formats for JSON, change tracking, and screenshots:

JSON format

 curl -X POST https://api.firecrawl.dev/v2/scrape \
     -H 'Content-Type: application/json' \
     -H 'Authorization: Bearer YOUR_API_KEY' \
     -d '{
       "url": "https://docs.firecrawl.dev/",
       "formats": [{
         "type": "json",
         "prompt": "Extract the company mission from the page."
       }]
     }'

Screenshot format

  curl -X POST https://api.firecrawl.dev/v2/scrape \
      -H 'Content-Type: application/json' \
      -H 'Authorization: Bearer YOUR_API_KEY' \
      -d '{
        "url": "https://docs.firecrawl.dev/",
        "formats": [{
          "type": "screenshot",
          "fullPage": true,
          "quality": 80,
          "viewport": { "width": 1280, "height": 800 }
        }]
      }'

Crawl options mapping (v1 → v2)

v1v2
allowBackwardCrawling(removed) use crawlEntireDomain
maxDepth(removed) use maxDiscoveryDepth
ignoreSitemap (bool)sitemap (e.g., "only", "skip", or "include")
(none)prompt

Crawl prompt + params preview

See crawl params preview examples:

  curl -X POST https://api.firecrawl.dev/v2/crawl-params-preview \
      -H 'Content-Type: application/json' \
      -H 'Authorization: Bearer YOUR_API_KEY' \
      -d '{
        "url": "https://docs.firecrawl.dev",
        "prompt": "Extract docs and blog"
      }'

What's Changed

New Contributors

Full Changelog: https://github.com/firecrawl/firecrawl/compare/v1.15.0...v2.0.0

Jul 18, 2025

Firecrawl v1.15.0 is here!

  • SSO for enterprise
  • Improved scraping reliability
  • Search params added to activity logs
  • FireGEO example
  • And over 50 PRs merged for bug & improvements 🔥

Improvements

  • OMCE support in scrapeURL and HTML transformer
  • Improved logging (search params, cache age)
  • New created_at field in /crawl/active response
  • Case-insensitive URL protocol checks
  • filterLinks ported to Rust

Fixes

  • HTML transformer stability (Arabic, base tag, panic)
  • scrapeURL index bug & waitFor exclusion
  • PDF billing bug with parsePDF=false
  • Crawl returning only 1 result (edge case)
  • Crawler no-sections bug
  • Logger method naming
  • Timeout handling in API
  • crawl-status resilience for ejected jobs

SDK & Infra

  • SDK header param fix & async error handling
  • Express port now configurable via env var
  • Temporary crawl expiry exemption

Docs

  • Kubernetes setup update

What's Changed

New Contributors

Full Changelog: https://github.com/mendableai/firecrawl/compare/v1.14.0...v1.15.0

Jul 4, 2025

v1.14.0 Release

We're excited to announce the release of Firecrawl v1.14.0, packed with cool updates.

  • Authenticated scraping (Join the waitlist here)
  • Zero data retention for enterprise (Email us at help@firecrawl.com to enable it)
  • Improved p75 speeds
  • New MCP version w/ maxAge + better tool calling
  • Open Researcher Example (Open Source Researcher). See repo
  • And so much more 🔥

What's Changed

Full Changelog: https://github.com/mendableai/firecrawl/compare/v1.13.0...v1.14.0

Jun 27, 2025

Firecrawl v1.13.0 Changelog

  • Added AU, FR, DE to Stealth Mode
  • Crawl subdomains with allowSubdomains
  • Google slides scraping
  • Generate a PDF of the current page. See docs
  • Higher res screenshots with quality param
  • Weekly view for usage on the dashboard
  • Fireplexity Example (Open Source Perplexity). See repo
  • And more!

Features

  • Support cookies in scrapeURL via safeFetch
  • Add Google Slides scraping
  • Add PDF actions to API, Python & JS SDKs
  • Screenshot quality config
  • parsePDF support in Python & JS SDKs
  • Allow international URLs
  • Enable subdomain crawling in SDKs
  • Worker liveness check before job queueing

Improvements & Fixes

  • Removed old cache systems (Redis, PDF cache, etc.)
  • Enhanced error messages (TLS, DNS)
  • Improved job count query & error handling
  • Fixed pdf, credits scope, ignoreInvalidURLs bugs
  • Reduced over-logging in concurrency limits
  • Added special job expiration bypass for Faire team
  • Unified fallback logic in JS SDK

What's Changed

New Contributors

Full Changelog: https://github.com/mendableai/firecrawl/compare/v1.12.0...v1.13.0

Jun 20, 2025

Firecrawl v1.12.0 is here!

  • New Concurrency System - Specify max concurrency by request for better control. See docs.
  • Crawl Entire Domain Param - Follow internal links to sibling or parent URLs, not just child paths (prev. allowBackwardLinks). See docs.
  • Google Docs Scraping - We now officially support scraping Google Docs files
  • Improved Activity Logs - Better support for FIRE-1 requests. See your logs here.
  • /search Playground Enhanced - Location Params added. Check out the playground.
  • Firestarter Example - Open Source Chatbot building platform. Repo here.
  • Plus tons of performance improvements and bug fixes. 

P.S. Have feedback or ideas for v1.13.0? Hit reply and let us know. We're always listening to our community to build the features you need most.

What's Changed

Full Changelog: https://github.com/mendableai/firecrawl/compare/v1.11.0...v1.12.0

Jun 13, 2025

Firecrawl v1.11.0 is here!

Major Features

  • Launched our Firecrawl Index
    • Speed up scrapes 5x if opted in
  • Improved Activity Logs
    • View webhook events
    • Active crawl management
  • Fire Enrich Example (Open Source Clay)
  • Community Java SDK
  • and a lot more

Features

  • Improved Playwright tests and webhook test coverage
  • Added GET /crawl/ongoing endpoint
  • Introduced tag support in change tracking
  • Added integration field to jobs and propagated through queue worker
  • Parallel testing for runpod v2 and updated mu
  • Ported queryIndexAtSplitLevel to RPC
  • Enhanced SDK with index and missing parameters
  • Removed redundant GCS check to improve performance
  • Added credits_billed field across pipeline
  • Enabled domain-level index splitting for better map querying
  • Used index in search and extract operations
  • Removed unused index columns

Fixes & Improvements

  • Fixed crawl pre-finishing logic
  • Refactored callWebhook and added logging
  • Improved index testing (FIR-2214)
  • Fixed JS SDK tests
  • Clarified scrape options usage in README
  • Fixed missing PLAYWRIGHT_MICROSERVICE_URL in env example
  • Improved concurrency limit notification emails
  • Removed query param sanitization that broke extract

Contributors

  • @Arkit003 made their first contribution
  • @pulbyte made their first contribution

Full Changelog: https://github.com/mendableai/firecrawl/compare/v1.10.0...v1.11.0

What's Changed

New Contributors

Full Changelog: https://github.com/mendableai/firecrawl/compare/v1.10.0...v1.11.0

Jun 3, 2025
Introducing Search - v1.10.0

We’re excited to announce the launch of our new Search API endpoint that combines web search with Firecrawl’s powerful scraping capabilities.

Search Features:

  • Search the web and get full content from results in one API call
  • Choose specific output formats (markdown, HTML, links, screenshots)
  • Customize search parameters (language, country, time range, number of results)
  • Full SDK support for Python and Node.js

More Features

  • Auto mode proxy for scraping (scrapeURL, js-sdk) #1551, #1602
  • Timeout handling and content type improvements for scrapeURL/pdf #1570, #1604, #1592
  • Redis improvements: separate non-eviction Redis support #1600
  • Search improvements: ignoreBlockedURLs, ignore concurrency limit #1580, #1617
  • New /cclog endpoint for concurrency logging #1589
  • Metadata extraction now includes itemprop attributes #1624
  • Self-hosted: deployable Playwright image #1625

Fixes & Improvements

  • Better subdomain handling for LLMs.txt + bypass option #1557
  • Improved URL validation and special character handling #1547
  • Zombie worker cleanup + TTL handling for extract status #1575, #1599
  • Fix concurrency queue logic and rate limiter override #1595, #1593
  • Better logging for search pagination and robust fetch #1572, #1588
  • Minor fixes: og:locale:alternate, adblock toggle, Playwright-only logic, malformed metadata arrays #1597, #1616, #1574

Testing & Docs

  • Add MAX_RAM and MAX_CPU environment variable docs #1581
  • Testing infrastructure improvements #1623

What's Changed

New Contributors

Full Changelog: https://github.com/mendableai/firecrawl/compare/v1.9.0...v.10.0

May 16, 2025

Firecrawl v1.9.0 Release

What's New:

Self-Host Improvements

  • Supabase client fixes
  • Fixed support for LLM Providers
  • Crawl is much faster now
  • Global adoption of cacheable lookup system
  • Easier setup

MCP Improvements (v1.11.0)

  • Tons of improvements to it (prompts, examples, and how to use params properly)

SDK & API Enhancements

  • Added change tracking to SDK 2.0
  • Crawl delay support with per-crawl concurrency limiting
  • New Qwen3 crawler example via OpenRouter
  • Cancel batch scrape endpoint

Performance & Limits

  • Global adoption of cacheable lookup system
  • Increased map endpoint limit from 5,000 to 30,000 links
  • Search schema limit increased from 50 to 100

Fixes & Stability

  • Better error handling for SSL failures
  • Optional chaining bug fixes
  • WaitAction field validation in firecrawl-py
  • Concurrency queue reworked to prioritize by time, not priority

Dashboard (Cloud version)

  • New activity logs

What's Changed

New Contributors

Full Changelog: https://github.com/mendableai/firecrawl/compare/v1.8.0...v1.9.0

Apr 28, 2025
Launch Week III - v.1.8.0

Day 7 – Integration Day

Overview

For the final day of Launch Week III, we’re rolling out new and updated integrations that make it easier to connect Firecrawl to the tools and platforms you already use.

From automation platforms to AI pipelines, Firecrawl now integrates with 20+ services, giving you a faster path from web data to workflow execution.

Key Integration Updates

  • Discord Bot: Trigger scrapes and receive structured results right inside your server. Learn more
  • Make Integration: Visually build workflows powered by Firecrawl's scraping and extraction. Learn more
  • n8n Integration: Connect Firecrawl to custom automation flows. Learn more
  • Langflow: Seamlessly embed Firecrawl agents into Langflow pipelines. Learn more
  • LlamaIndex: Use Firecrawl with LlamaIndex to enrich and retrieve data intelligently. Learn more
  • Dify: Integrate Firecrawl with Dify to automate AI workflows. Learn more

More integrations are on the way — and if there’s one you’re missing, we’d love to hear about it.

Day 6 – Firecrawl MCP

Overview

Today we’re launching a major upgrade to our Firecrawl MCP server, our implementation of the Model Context Protocol for LLM-connected scraping workflows.

This release brings FIRE-1 agent support to the MCP, letting you unlock data hidden behind interaction barriers like logins and buttons — all via scrape and extract endpoints.

We’re also introducing Server-Sent Events (SSE) support for local use, making setup and real-time integration easier than ever.

These updates make it simpler to stream web data into LLM pipelines, with intelligent agents handling the heavy lifting.

Key Features

  • FIRE-1 Support: Use the web action agent via MCP’s scrape and extract endpoints.
  • Interaction-Aware: Automatically navigates complex web barriers.
  • SSE Support: Stream output in real time — locally or from our hosted MCP server.
  • Plug-and-Play: Minimal setup needed for integration with your stack.

Day 5 – Developer Day

Overview

Today is all about developers. We’re rolling out upgrades that make building with Firecrawl smoother and more scalable — whether you’re working in Python, Rust, or your favorite editor.

We’ve introduced a fully async Python SDK with named params and return types, powerful new features in the Rust SDK, expanded team support on every plan, and a brand new Firecrawl Dark Theme for VSCode and compatible editors.

Key Features

  • Major Python SDK Improvements: Named params, return types, and Async added to our Python SDK (VERSION 2.0)
  • Rust SDK Upgrades: Batch scraping, job cancellation, llms.txt generation, smarter search.
  • Firecrawl Dark Theme: A sleek new theme for VSCode, Cursor, and Windsurf. Download it here.
  • QoL Improvements: Plenty of small but meaningful changes to improve dev experience like no more team caps. All plans now include up to 20 seats.

Day 4 – LLMstxt.new

Overview

Today we’re announcing http://llmstxt.new — the fastest way to turn any website into a clean, consolidated text file for LLMs.

Just add llmstxt.new/ in front of any URL, and you’ll get back a plain .txt file, optimized for AI training and inference. No boilerplate, no noise — just useful content.

Built on top of Firecrawl, this tool makes it effortless to prepare real-world web content for use in LLM pipelines.

Key Features

  • Instant Usage: Just add llmstxt.new/ before any URL.
  • Two Outputs: llms.txt for concise summaries, llms-full.txt for full content.
  • API Friendly: Call via http://llmstxt.new/{YOUR_URL} or with a Firecrawl API key for full output.
  • LLM-Ready Format: Designed for both training and inference contexts.

Day 3 – /extract v2

Overview

Today we’re launching /extract v2, a major upgrade to our extraction system — powered by the FIRE-1 agent.

With full support for pagination, multi-step flows, and dynamic interactions, extract v2 goes way beyond what we shipped back in January. It’s also now possible to extract data without a URL, using a built-in search layer to find the content you’re after.

We’ve rebuilt the internals from the ground up — improved models, better architecture, and significantly better performance across our internal benchmarks.

Key Features

  • FIRE-1 Integration: Automatically handles page actions, navigation, and interaction.
  • Multi-Page Extraction: Extract data across paginated or step-based flows.
  • No URL Required: Use built-in search to extract content without a direct link.
  • Improved Accuracy: New models and architecture mean better, faster results.

Day 2 - FIRE-1 Agent

Overview

Meet FIRE-1, Firecrawl's first AI Agent built to take web scraping to the next level. With intelligent navigation and interaction capabilities, FIRE-1 can go far beyond traditional scraping methods.

From handling pagination to interacting with dynamic site elements like buttons and links, FIRE-1 allows for powerful, context-aware scraping and extraction workflows.

Key Features

  • Smart Navigation: Move through paginated or multi-step content automatically.
  • Dynamic Interaction: Click buttons, follow links, and interact with JavaScript-rendered components.

Day 1 - Change Tracking

Overview

Change tracking is a powerful feature that allows you to monitor and detect changes in web content over time. It is available in both the JavaScript and Python SDKs.

Key Features

  • Detect Changes: Identify if a webpage has changed since the last scrape.
  • View Specific Changes: Access detailed information about what has changed between scrapes.
  • Structured Data: Receive structured data about the changes.
  • Visibility Control: Manage the visibility of changes on your website.

Using the changeTracking format, you can effectively monitor changes on a website and receive comprehensive information about the timestamp of the previous scrape, the result of the comparison between the two page versions, and the visibility of the current page/URL.

Day 0 - Firecrawl Editor Theme

Overview

We're excited to release our official Firecrawl Editor Theme! Available now for most editors including Cursor, Windsurf, and more.

Key Features

The Firecrawl Editor Theme provides a clean, focused coding experience for everyone. Our color palette emphasizes readability while maintaining the Firecrawl brand identity.

Download

You can download the editor theme on the VS Code Marketplace here.

What's Changed

New Contributors

Full Changelog: https://github.com/mendableai/firecrawl/compare/v1.7.0...v1.8.0

Apr 4, 2025

v1.7.0 - Release Notes

New Features

  • Deep Research Open Alpha: Structured outputs + customizability.
  • llmstxt.new: Generate an llms.txt for any website by just appending its url: llmstxt.new/firecrawl.dev
  • Concurrent Browsers: Improved rate limits for all users.
  • Compare Beta: Figure what has changed in the website directly in /scrape and /crawl endpoints. Currently in closed beta.
  • /extract: URLs are now optional.
  • /scrape: Warns if concurrency-limited.
  • New Firecrawl Examples: Featuring models like Claude 3.7, Gemini 2.5, Deepseek V3, Mistral 3.1, and more.
  • Crawl: maxDiscoveryDepth option added.

Fixes & Improvements

  • Fixed circular JSON error in search.
  • Reworked new tally system.
  • Fixed sitemaps poisoning crawler with unrelated links.
  • Crawler status retries added on failure (up to 3 times).
  • Credit check now snaps to remaining credits if exceeded.
  • Fixed path filtering bug in Map.
  • Removed unsupported schema in llmExtract.

What's Changed

New Contributors

Full Changelog: https://github.com/mendableai/firecrawl/compare/v1.6.0...v1.7.0

Previous12Next
Latest
v2.9.0
Tracking Since
Sep 5, 2024
Last fetched Apr 18, 2026