AI & Machine Learning

Claude Opus 4.7: What Developers Actually Need to Know

Anthropic released Claude Opus 4.7 on April 16, 2026. A hands-on look at what changed, how it compares to Opus 4.6, the new xhigh effort level and task budgets, and what the Project Glasswing safeguards mean for production AI systems.

Harsh RastogiHarsh Rastogi
Apr 19, 202610 min
ClaudeAnthropicAI ModelsLLMAgentic AIAI & Machine Learning

TL;DR

Claude Opus 4.7 (model ID claude-opus-4-7) is Anthropic's newest flagship model, released on April 16, 2026. Pricing is unchanged at $5 / $25 per million input/output tokens, but the model adds self-verification on coding tasks, 3x higher image resolution, a new "xhigh" reasoning effort level, task budgets in public beta, and Project Glasswing cyber safeguards. For anyone building production AI or agentic coding workflows, this is the strongest generally available model on the market today.

Why This Release Matters

Anthropic just shipped Claude Opus 4.7. It is a direct upgrade to Opus 4.6 and drops into the Claude API, Amazon Bedrock, Google Cloud Vertex AI, and Microsoft Foundry without code changes beyond the model string.

The headline is not "higher benchmark scores", it is reliability on long-running tasks. Opus 4.7 was tuned for engineers who hand off multi-step work to an agent and walk away. That one shift changes what you can delegate in production, which is why this release lands heavier than the version number suggests.

Claude Opus 4.7 vs Claude Opus 4.6

FeatureClaude Opus 4.6Claude Opus 4.7
Release dateFeb 2026April 16, 2026
Input pricing$5 / MTok$5 / MTok
Output pricing$25 / MTok$25 / MTok
Max image resolution~784 px2,576 px (3x)
Effort levelslow, medium, high, max+ new "xhigh"
Task budgetsNoYes (public beta)
Self-verificationLimitedNative
Cyber safeguardsBasicProject Glasswing
Coding benchmark deltaBaseline+13% resolution

What Actually Changed

1. Self-verification on coding tasks

Opus 4.7 catches its own logical faults during planning, verifies outputs before reporting back, and resists the pattern of generating plausible-but-wrong fallbacks that plagued older models. The effect shows up in numbers: early benchmarks show a 13% resolution improvement over Opus 4.6 on a 93-task coding benchmark, including four tasks that neither Opus 4.6 nor Sonnet 4.6 could solve.

For production code, this changes the trust equation. With Opus 4.6 I reviewed every line of AI-generated code on critical paths. With 4.7, the model is doing more of that review itself, and doing it well enough that the human checkpoint can move one layer up the stack.

2. Vision at 3x resolution

Claude Opus 4.7 accepts images up to 2,576 pixels on the long edge, more than three times the limit of earlier Claude models. That unlocks:

  • Reading dense technical diagrams without downscaling
  • Processing 4K screenshots end-to-end
  • Analyzing architectural drawings and blueprints
  • Extracting information from complex UI mockups and Figma exports
  • Parsing high-density dashboards, financial charts, and stack traces

For any workflow where you feed Claude screenshots of logs, dashboards, or architecture diagrams, this is a large quality-of-life jump.

3. The new "xhigh" effort level

Anthropic added a new reasoning effort tier called "xhigh", sitting between "high" and "max". The point is granular control over the tradeoff between reasoning depth and latency. For agentic coding, Anthropic recommends starting with "high" or "xhigh". In practice, most tasks that used to need "max" can now run on "xhigh" with lower latency and cost, and "max" is reserved for the genuinely hard problems.

4. Task budgets (public beta)

Task budgets let you cap how many reasoning tokens Claude consumes on a given task, via a budget_tokens parameter on the thinking config. For anyone running dozens of concurrent agents in CI, automated code review, or data-processing pipelines, this is the cost-control lever teams have been waiting on. Agent costs become deterministic instead of a surprise at the end of the month.

5. Project Glasswing cyber safeguards

Opus 4.7 is Anthropic's first model with built-in cyber safeguards that automatically detect and block prohibited or high-risk cybersecurity uses. During training, Anthropic traded off some cyber capability for safety. This matters because it is tied to Claude Mythos Preview, Anthropic's more powerful model that is not publicly available due to cyber concerns. Opus 4.7 is the testing ground for the safeguards that would eventually let Mythos-class models ship broadly.

What This Means for Production AI Systems

I build production Generative AI systems at Modelia.ai, and previously built an Agentic AI interviewer at Asynq.ai. Here is what Opus 4.7 changes in practice.

Agentic workflows get more reliable

Self-verification is the most load-bearing change for anyone running AI agents in production. When an agent is making decisions autonomously, for example processing resumes, generating images, executing multi-step fashion workflows, or evaluating candidates, the model catching its own mistakes before acting on them drops the error rate that previously forced human checkpoints in the middle of the flow.

That does not eliminate human-in-the-loop for high-stakes actions. It does make the "autonomous within bounds" pattern practical for a lot more of the workflow.

The cost equation shifts

Task budgets plus the xhigh effort level equal predictable cost for agentic workloads. Before Opus 4.7, you either overspent on "max" everywhere or under-invested on "high" and got inconsistent results. Now you can tune per task type:

  • Code generation: xhigh effort with a task budget
  • Code review: high effort, faster and cheaper
  • Complex debugging: max effort, reserved for hard cases
  • Simple edits and formatting: medium effort
  • Classification and routing: low effort

Plan for the tokenizer change

One thing to watch: Opus 4.7 uses an updated tokenizer. The same input can map to roughly 1.0 to 1.35x more tokens depending on content type. Per-token pricing did not change, but token counts for the same payloads did, so monitor your spend after switching. A 15 to 35 percent effective cost bump on the same inputs is realistic for some workloads.

Migration Checklist: Opus 4.6 to Opus 4.7

If you are on claude-opus-4-6, here is the playbook:

  • Change claude-opus-4-6 to claude-opus-4-7 in your API calls, environment variables, and SDK configs.
  • Re-test prompts. Opus 4.7 is more precise on instruction following, so prompts that Opus 4.6 interpreted generously may need to be more explicit.
  • Monitor token usage. The new tokenizer can map the same input to up to 1.35x more tokens.
  • Try "xhigh" on your hardest tasks. If you were using "max" everywhere, downgrade what you can to "xhigh" and keep "max" for the genuinely hard problems.
  • Add budget_tokens on long-running agents. This is the single biggest cost-control lever in this release.
  • Plan the deprecation. Opus 4.6 is being deprecated, and Claude Sonnet 4 and Claude Opus 4 retire on June 15, 2026. Migrate before then, not on the last day.

The Mythos Shadow

The most interesting subtext of this release is what it says about Claude Mythos Preview. Anthropic is being openly transparent: Opus 4.7 is good, but Mythos is better, and they are not releasing Mythos because the cyber implications are too high. Read that as responsible AI development, strategic positioning, or both. Either way, it tells you the frontier is meaningfully ahead of what ships today. The safeguards being tested on Opus 4.7 are practice for the eventual broader release of Mythos-class models.

The pragmatic takeaway: Claude Opus 4.7 is the best generally available AI model right now, so use it. But build your systems model-agnostic. The upgrade cycle is accelerating (Opus 4.5 to 4.6 to 4.7, each roughly two months apart), and your architecture should handle model swaps without a rewrite.

Claude Opus 4.7 FAQ

Is Claude Opus 4.7 better than GPT-4 Turbo or Gemini for coding?

On agentic coding benchmarks, Claude Opus 4.7 is currently ahead. Self-verification and tool-use reliability put it in front for multi-step engineering workflows, which is where most production coding agents actually live.

How much does Claude Opus 4.7 cost?

$5 per million input tokens and $25 per million output tokens, same as Opus 4.6. The updated tokenizer can produce 1.0 to 1.35x more tokens for the same input, so effective spend may rise slightly.

When is Claude Opus 4.6 being deprecated?

Opus 4.6 is being deprecated, and Claude Sonnet 4 and Claude Opus 4 retire on June 15, 2026. Migrate to claude-opus-4-7 before then.

What is the xhigh effort level?

A new reasoning effort tier between "high" and "max". It gives you near-max quality at lower latency and cost, which suits agentic coding and long-running agents.

What are task budgets?

A public-beta parameter (budget_tokens) that caps the reasoning tokens Claude can spend on a given task. It makes agent costs deterministic and prevents runaway spending on hard tasks.

Does Claude Opus 4.7 support tool use and MCP?

Yes. Tool use is fully supported and more reliable than Opus 4.6. The model works with Anthropic's Model Context Protocol (MCP) and existing Anthropic SDK integrations.

Where can I use Claude Opus 4.7?

On the Claude API, Amazon Bedrock, Google Cloud Vertex AI, Microsoft Foundry, and Claude Code, Anthropic's official CLI for engineers.

Bottom Line

Claude Opus 4.7 is not a revolution, it is a practical upgrade that makes AI-assisted engineering more reliable and more controllable. Self-verification, task budgets, and the xhigh effort level are the features that show up in day-to-day engineering work, which is exactly where it matters.

If you build with Claude in production, upgrade now. If you are evaluating AI models for an engineering workflow, Opus 4.7 is the strongest generally available option on the market today.

The more interesting story is the pace. Two-month upgrade cycles, each release meaningfully better than the last. Compound that over 12 months and what is possible in software engineering looks noticeably different.

Share this article

Harsh Rastogi - Full Stack Engineer

Harsh Rastogi

Full Stack Engineer

Full Stack Engineer building production AI systems at Modelia. Previously at Asynq and Bharat Electronics Limited. Published researcher.

Connect on LinkedIn

Follow me for more insights on software engineering, system design, and career growth.

View Profile