TL;DR
Claude Opus 4.7 (model ID claude-opus-4-7) is Anthropic's newest flagship model, released on April 16, 2026. Pricing is unchanged at $5 / $25 per million input/output tokens, but the model adds self-verification on coding tasks, 3x higher image resolution, a new "xhigh" reasoning effort level, task budgets in public beta, and Project Glasswing cyber safeguards. For anyone building production AI or agentic coding workflows, this is the strongest generally available model on the market today.
Why This Release Matters
Anthropic just shipped Claude Opus 4.7. It is a direct upgrade to Opus 4.6 and drops into the Claude API, Amazon Bedrock, Google Cloud Vertex AI, and Microsoft Foundry without code changes beyond the model string.
The headline is not "higher benchmark scores", it is reliability on long-running tasks. Opus 4.7 was tuned for engineers who hand off multi-step work to an agent and walk away. That one shift changes what you can delegate in production, which is why this release lands heavier than the version number suggests.
Claude Opus 4.7 vs Claude Opus 4.6
| Feature | Claude Opus 4.6 | Claude Opus 4.7 |
|---|---|---|
| Release date | Feb 2026 | April 16, 2026 |
| Input pricing | $5 / MTok | $5 / MTok |
| Output pricing | $25 / MTok | $25 / MTok |
| Max image resolution | ~784 px | 2,576 px (3x) |
| Effort levels | low, medium, high, max | + new "xhigh" |
| Task budgets | No | Yes (public beta) |
| Self-verification | Limited | Native |
| Cyber safeguards | Basic | Project Glasswing |
| Coding benchmark delta | Baseline | +13% resolution |
What Actually Changed
1. Self-verification on coding tasks
Opus 4.7 catches its own logical faults during planning, verifies outputs before reporting back, and resists the pattern of generating plausible-but-wrong fallbacks that plagued older models. The effect shows up in numbers: early benchmarks show a 13% resolution improvement over Opus 4.6 on a 93-task coding benchmark, including four tasks that neither Opus 4.6 nor Sonnet 4.6 could solve.
For production code, this changes the trust equation. With Opus 4.6 I reviewed every line of AI-generated code on critical paths. With 4.7, the model is doing more of that review itself, and doing it well enough that the human checkpoint can move one layer up the stack.
2. Vision at 3x resolution
Claude Opus 4.7 accepts images up to 2,576 pixels on the long edge, more than three times the limit of earlier Claude models. That unlocks:
- ›Reading dense technical diagrams without downscaling
- ›Processing 4K screenshots end-to-end
- ›Analyzing architectural drawings and blueprints
- ›Extracting information from complex UI mockups and Figma exports
- ›Parsing high-density dashboards, financial charts, and stack traces
For any workflow where you feed Claude screenshots of logs, dashboards, or architecture diagrams, this is a large quality-of-life jump.
3. The new "xhigh" effort level
Anthropic added a new reasoning effort tier called "xhigh", sitting between "high" and "max". The point is granular control over the tradeoff between reasoning depth and latency. For agentic coding, Anthropic recommends starting with "high" or "xhigh". In practice, most tasks that used to need "max" can now run on "xhigh" with lower latency and cost, and "max" is reserved for the genuinely hard problems.
4. Task budgets (public beta)
Task budgets let you cap how many reasoning tokens Claude consumes on a given task, via a budget_tokens parameter on the thinking config. For anyone running dozens of concurrent agents in CI, automated code review, or data-processing pipelines, this is the cost-control lever teams have been waiting on. Agent costs become deterministic instead of a surprise at the end of the month.
5. Project Glasswing cyber safeguards
Opus 4.7 is Anthropic's first model with built-in cyber safeguards that automatically detect and block prohibited or high-risk cybersecurity uses. During training, Anthropic traded off some cyber capability for safety. This matters because it is tied to Claude Mythos Preview, Anthropic's more powerful model that is not publicly available due to cyber concerns. Opus 4.7 is the testing ground for the safeguards that would eventually let Mythos-class models ship broadly.
What This Means for Production AI Systems
I build production Generative AI systems at Modelia.ai, and previously built an Agentic AI interviewer at Asynq.ai. Here is what Opus 4.7 changes in practice.
Agentic workflows get more reliable
Self-verification is the most load-bearing change for anyone running AI agents in production. When an agent is making decisions autonomously, for example processing resumes, generating images, executing multi-step fashion workflows, or evaluating candidates, the model catching its own mistakes before acting on them drops the error rate that previously forced human checkpoints in the middle of the flow.
That does not eliminate human-in-the-loop for high-stakes actions. It does make the "autonomous within bounds" pattern practical for a lot more of the workflow.
The cost equation shifts
Task budgets plus the xhigh effort level equal predictable cost for agentic workloads. Before Opus 4.7, you either overspent on "max" everywhere or under-invested on "high" and got inconsistent results. Now you can tune per task type:
- ›Code generation: xhigh effort with a task budget
- ›Code review: high effort, faster and cheaper
- ›Complex debugging: max effort, reserved for hard cases
- ›Simple edits and formatting: medium effort
- ›Classification and routing: low effort
Plan for the tokenizer change
One thing to watch: Opus 4.7 uses an updated tokenizer. The same input can map to roughly 1.0 to 1.35x more tokens depending on content type. Per-token pricing did not change, but token counts for the same payloads did, so monitor your spend after switching. A 15 to 35 percent effective cost bump on the same inputs is realistic for some workloads.
Migration Checklist: Opus 4.6 to Opus 4.7
If you are on claude-opus-4-6, here is the playbook:
- ›Change
claude-opus-4-6toclaude-opus-4-7in your API calls, environment variables, and SDK configs. - ›Re-test prompts. Opus 4.7 is more precise on instruction following, so prompts that Opus 4.6 interpreted generously may need to be more explicit.
- ›Monitor token usage. The new tokenizer can map the same input to up to 1.35x more tokens.
- ›Try "xhigh" on your hardest tasks. If you were using "max" everywhere, downgrade what you can to "xhigh" and keep "max" for the genuinely hard problems.
- ›Add
budget_tokenson long-running agents. This is the single biggest cost-control lever in this release. - ›Plan the deprecation. Opus 4.6 is being deprecated, and Claude Sonnet 4 and Claude Opus 4 retire on June 15, 2026. Migrate before then, not on the last day.
The Mythos Shadow
The most interesting subtext of this release is what it says about Claude Mythos Preview. Anthropic is being openly transparent: Opus 4.7 is good, but Mythos is better, and they are not releasing Mythos because the cyber implications are too high. Read that as responsible AI development, strategic positioning, or both. Either way, it tells you the frontier is meaningfully ahead of what ships today. The safeguards being tested on Opus 4.7 are practice for the eventual broader release of Mythos-class models.
The pragmatic takeaway: Claude Opus 4.7 is the best generally available AI model right now, so use it. But build your systems model-agnostic. The upgrade cycle is accelerating (Opus 4.5 to 4.6 to 4.7, each roughly two months apart), and your architecture should handle model swaps without a rewrite.
Claude Opus 4.7 FAQ
Is Claude Opus 4.7 better than GPT-4 Turbo or Gemini for coding?
On agentic coding benchmarks, Claude Opus 4.7 is currently ahead. Self-verification and tool-use reliability put it in front for multi-step engineering workflows, which is where most production coding agents actually live.
How much does Claude Opus 4.7 cost?
$5 per million input tokens and $25 per million output tokens, same as Opus 4.6. The updated tokenizer can produce 1.0 to 1.35x more tokens for the same input, so effective spend may rise slightly.
When is Claude Opus 4.6 being deprecated?
Opus 4.6 is being deprecated, and Claude Sonnet 4 and Claude Opus 4 retire on June 15, 2026. Migrate to claude-opus-4-7 before then.
What is the xhigh effort level?
A new reasoning effort tier between "high" and "max". It gives you near-max quality at lower latency and cost, which suits agentic coding and long-running agents.
What are task budgets?
A public-beta parameter (budget_tokens) that caps the reasoning tokens Claude can spend on a given task. It makes agent costs deterministic and prevents runaway spending on hard tasks.
Does Claude Opus 4.7 support tool use and MCP?
Yes. Tool use is fully supported and more reliable than Opus 4.6. The model works with Anthropic's Model Context Protocol (MCP) and existing Anthropic SDK integrations.
Where can I use Claude Opus 4.7?
On the Claude API, Amazon Bedrock, Google Cloud Vertex AI, Microsoft Foundry, and Claude Code, Anthropic's official CLI for engineers.
Bottom Line
Claude Opus 4.7 is not a revolution, it is a practical upgrade that makes AI-assisted engineering more reliable and more controllable. Self-verification, task budgets, and the xhigh effort level are the features that show up in day-to-day engineering work, which is exactly where it matters.
If you build with Claude in production, upgrade now. If you are evaluating AI models for an engineering workflow, Opus 4.7 is the strongest generally available option on the market today.
The more interesting story is the pace. Two-month upgrade cycles, each release meaningfully better than the last. Compound that over 12 months and what is possible in software engineering looks noticeably different.
