Launched Claude Opus 4.7, the most capable generally available model for complex reasoning and agentic coding, at the same $5 / $25 per MTok pricing as Opus 4.6. Opus 4.7 includes API breaking changes versus Opus 4.6; see migration guide before upgrading. Includes capability improvements, new features, and updated tokenizer.
We announced the deprecation of the Claude Sonnet 4 model (claude-sonnet-4-20250514) and the Claude Opus 4 model (claude-opus-4-20250514), with retirement on the Claude API scheduled for June 15, 2026. We recommend migrating to Claude Sonnet 4.6 and Claude Opus 4.6 respectively.
Launched the advisor tool in public beta. Pair a faster executor model with a higher-intelligence advisor model that provides strategic guidance mid-generation. Long-horizon agentic workloads get close to advisor-solo quality while the bulk of token generation happens at executor-model rates. Include the beta header advisor-tool-2026-03-01 in your requests.
Launched Claude Managed Agents in public beta, a fully managed agent harness for running Claude as an autonomous agent with secure sandboxing, built-in tools, and server-sent event streaming. Create agents, configure containers, and run sessions through the API. All endpoints require the managed-agents-2026-04-01 beta header.
Launched the ant CLI, a command-line client for the Claude API that enables faster interaction with the Claude API, native integration with Claude Code, and versioning of API resources in YAML files.
Announced Claude Mythos Preview available as a gated research preview for defensive cybersecurity work as part of Project Glasswing. Access is invitation-only.
The Messages API is now available on Amazon Bedrock as a research preview. The new Claude in Amazon Bedrock endpoint at /anthropic/v1/messages uses the same request shape as the first-party Claude API and runs on AWS-managed infrastructure. Available in us-east-1; contact your Anthropic account executive to request access.
Raised the max_tokens cap to 300k on the Message Batches API for Claude Opus 4.6 and Sonnet 4.6. Include the output-300k-2026-03-24 beta header to generate longer single-turn outputs.
Retiring the 1M token context window beta for Claude Sonnet 4.5 and Claude Sonnet 4 on April 30, 2026. After that date, the context-1m-2025-08-07 beta header will have no effect on these models, and requests that exceed the standard 200k-token context window will return an error. Migrate to Claude Sonnet 4.6 or Claude Opus 4.6, which support the full 1M token context window at standard pricing with no beta header required.
Added model capability fields to the Models API. GET /v1/models and GET /v1/models/{model_id} now return max_input_tokens, max_tokens, and a capabilities object. Query the API to discover what each model supports.
Launched the display field for extended thinking, letting you omit thinking content from responses for faster streaming. Set thinking.display: "omitted" to receive thinking blocks with an empty thinking field and the signature preserved for multi-turn continuity. Billing is unchanged.
The 1M token context window is now generally available for Claude Opus 4.6 and Sonnet 4.6 at standard pricing. Requests over 200k tokens work automatically for these models with no beta header required. The 1M token context window remains in beta for Claude Sonnet 4.5 and Sonnet 4.
Removed the dedicated 1M rate limits for all supported models. Your standard account limits now apply across every context length.
Raised the media limit from 100 to 600 images or PDF pages per request when using the 1M token context window.
Launched automatic caching for the Messages API. Add a single cache_control field to your request body and the system automatically caches the last cacheable block, moving the cache point forward as conversations grow. No manual breakpoint management required. Works alongside existing block-level cache control for fine-grained optimization. Available on the Claude API and Azure AI Foundry (preview).
Retired the Claude Sonnet 3.7 model (claude-3-7-sonnet-20250219) and the Claude Haiku 3.5 model (claude-3-5-haiku-20241022). All requests to these models will now return an error. Announced the deprecation of the Claude Haiku 3 model (claude-3-haiku-20240307), with retirement scheduled for April 19, 2026.
Launched Claude Sonnet 4.6, our latest balanced model combining speed and intelligence. Sonnet 4.6 delivers improved agentic search performance while consuming fewer tokens. Supports extended thinking and a 1M token context window (beta).
API code execution is now free when used with web search or web fetch.
The web search tool and programmatic tool calling are now generally available (no beta header required). Web search and web fetch now support dynamic filtering.
The code execution tool, web fetch tool, tool search tool, tool use examples, and memory tool are now generally available (no beta header required).
We've launched Claude Opus 4.6, our most intelligent model for complex agentic tasks and long-horizon work. Opus 4.6 recommends adaptive thinking (thinking: {type: "adaptive"}); manual thinking (type: "enabled" with budget_tokens) is deprecated. The effort parameter is now generally available and supports Claude Opus 4.6. We've launched the compaction API in beta, providing server-side context summarization for effectively infinite conversations. Introduced data residency controls with the inference_geo parameter for US-only inference at 1.1x pricing. The 1M token context window is now available in beta for Claude Opus 4.6. Fine-grained tool streaming is now generally available on all models.
Launched Claude Opus 4.6, our most intelligent model for complex agentic tasks and long-horizon work. Opus 4.6 recommends adaptive thinking (thinking: {type: "adaptive"}); manual thinking (type: "enabled" with budget_tokens) is deprecated. Opus 4.6 does not support prefilling assistant messages.
The effort parameter is now generally available (no beta header required) and supports Claude Opus 4.6.
Launched the compaction API in beta, providing server-side context summarization for effectively infinite conversations.
Introduced data residency controls, allowing you to specify where model inference runs with the inference_geo parameter. US-only inference available at 1.1x pricing.
The 1M token context window is now available in beta for Claude Opus 4.6. Long context pricing applies to requests exceeding 200k input tokens.
Fine-grained tool streaming is now generally available on all models and platforms (no beta header required). The output_format parameter for structured outputs has been moved to output_config.format.
Structured outputs are now generally available on the Claude API for Claude Sonnet 4.5, Claude Opus 4.5, and Claude Haiku 4.5. GA includes expanded schema support, improved grammar compilation latency, and simplified integration with no beta header required. The output_format parameter has moved to output_config.format.
console.anthropic.com now redirects to platform.claude.com. The Claude Console has moved to its new home as part of Claude brand consolidation. Existing bookmarks and links will continue working via automatic redirect.
We've retired the Claude Opus 3 model (claude-3-opus-20240229). All requests to this model will now return an error. We recommend upgrading to Claude Opus 4.5, which offers significantly improved intelligence at a third of the cost.