Claude logo on a laptop screen
News

Anthropic Launches Claude Opus 4.7 With Gains in Coding and Vision

3 minute read
Michelle Hawley avatar
By
SAVED
The latest Claude Opus model is built to handle longer, harder tasks — and block more risky cyber requests.

Anthropic has released Claude Opus 4.7, an upgrade to its flagship Opus model line that the company says delivers meaningful improvements in software engineering, visual understanding and long-running agentic tasks. The model debut also included a novel approach to managing AI's cybersecurity risks.

Table of Contents

A Coding Model You Can Actually Hand Off Work To

The biggest improvement with Opus 4.7 is in software engineering. According to Anthropic, users in early testing found they could delegate their most demanding, supervision-heavy coding tasks to the new model with confidence, something they couldn't reliably do with Opus 4.6.

That's a notable claim. Large language models (LLMs) have long struggled with what researchers call "long-horizon" tasks: multi-step jobs that require maintaining consistency and catching errors across many turns without human check-ins. Opus 4.7 reportedly does exactly that, with the model now verifying its own outputs before reporting results back to users.

The model also shows state-of-the-art performance on SWE-bench, a widely used software engineering benchmark, and tops leaderboard results on SWE-bench Multilingual and SWE-bench Multimodal variants.

Claude Opus benchmarks chart

What Else Is New With Opus 4.7

Better Eyes

Opus 4.7 accepts images up to 2,576 pixels on the long edge — roughly 3.75 megapixels — more than 3x the resolution supported by previous Claude models. That's a model-level change, meaning images are automatically processed at higher fidelity without any change to how developers send them.

The practical upside: use cases that depend on fine visual detail, like reading dense screenshots in computer-use agents or extracting data from complex diagrams, become substantially more viable.

Note: Higher-resolution images consume more tokens. Developers who don't need the extra detail can downsample images before sending them.

Sharper Instruction-Following — With a Catch

Anthropic describes Opus 4.7 as "substantially better" at following instructions. But the company flags an interesting side effect: prompts written for older models may now produce unexpected results, because earlier versions were prone to interpreting instructions loosely or skipping steps. Opus 4.7 takes instructions literally.

The company suggests that those migrating from Opus 4.6 or other older models should re-tune their prompts accordingly.

Better Memory Across Sessions

The model has improved ability to use file system-based memory, remembering important context across multi-session tasks. The result, according to Anthropic, is that later tasks in a long project require less up-front context-setting.

A New Approach to Cybersecurity Risk

The most distinctive feature of this release may be a safety mechanism.

Last week, Anthropic announced Project Glasswing, a new initiative acknowledging the risks and potential benefits of powerful AI for cybersecurity. As part of that project, the company committed to keeping its most powerful model, Claude Mythos Preview, on a restricted release while testing new cyber safeguards on less capable models first.

Opus 4.7 is that test case.

Learning Opportunities

The model includes automated systems that detect and block requests tied to prohibited or high-risk cybersecurity uses. Anthropic noted that during training, it also experimented with differentially reducing the model's cyber capabilities compared to Mythos Preview — an early example of targeted capability shaping, where developers try to make a model less capable in specific dangerous domains without degrading overall performance.

For legitimate security professionals — penetration testers, vulnerability researchers, red teamers — Anthropic is launching a Cyber Verification Program to grant access to capabilities that would otherwise be blocked.

Pricing, Availability & Migration Notes

DetailInfo
AvailabilityAll Claude products, API, Amazon Bedrock, Google Cloud Vertex AI, Microsoft Foundry
API model stringclaude-opus-4-7
Pricing $5/ million input tokens | $25/ million output tokens
TokenizerUpdated — same input may use 1.0-1.35x more tokens than Opus 4.6

The token increase is worth planning for. Opus 4.7 uses an updated tokenizer and also thinks more at higher effort levels, particularly in later turns of agentic sessions. Anthropic says the net effect is favorable on its internal coding benchmarks, but recommends teams measure real-world token usage on their own workloads before fully migrating.

Also Launching Today

  • xhigh effort level: A new setting between "high" and "max" for finer control over reasoning depth vs. latency on hard problems. Claude Code's default effort has been raised to xhigh for all plans.
  • Task budgets (public beta): A way for API developers to guide Claude's token spending across longer runs.
  • /ultrareview in Claude Code: A dedicated review session that audits code changes for bugs and design issues. Pro and Max users get three free trials.
  • Auto mode for Claude Code Max: Lets Claude make permission decisions on your behalf for longer, lower-interruption task runs.

According to Anthropic, Opus 4.7 shows a broadly similar safety profile to Opus 4.6, with improvements on honesty and resistance to prompt injection, and modest weaknesses in harm-reduction advice around controlled substances. Claude Mythos Preview remains the company's best-aligned model by its own evaluations.

About the Author
Michelle Hawley

Michelle Hawley is an experienced journalist who specializes in reporting on the impact of technology on society. As editorial director at Simpler Media Group, she oversees the day-to-day operations of VKTR, covering the world of enterprise AI and managing a network of contributing writers. She's also the host of CMSWire's CMO Circle and co-host of CMSWire's CX Decoded. With an MFA in creative writing and background in both news and marketing, she offers unique insights on the topics of tech disruption, corporate responsibility, changing AI legislation and more. She currently resides in Pennsylvania with her husband and two dogs. Connect with Michelle Hawley:

Main image: gguy | Adobe Stock
Featured Research