← Back to feed
8

Anthropic Releases Claude Opus 4.7: Hybrid Reasoning, 1M Context Window, and Cybersecurity Safeguards

ModelsTop News8 sources·Apr 16

Summary

  • • Claude Opus 4.7 launches as hybrid reasoning model with 1M token context window
  • • Adaptive thinking automatically scales compute based on task complexity
  • • 13% coding benchmark gain over Opus 4.6; first model with Project Glasswing cybersecurity safeguards
  • • New tokenizer generates 32–45% more tokens for identical text; OpenRouter analysis finds 12–27% effective cost increases despite unchanged per-token rates
Adjust signal

Details

1.Product Launch

Claude Opus 4.7 released as hybrid reasoning model with adaptive thinking and 1M token context window

Generally available across API and consumer tiers (Pro, Max, Team, Enterprise). The 1M token context window enables multi-day, multi-session enterprise and agentic workflows. Adaptive thinking automatically scales reasoning depth based on task complexity.

2.Stat

13% coding benchmark improvement over Opus 4.6 on a 93-task evaluation suite

Opus 4.7 solved four tasks that neither Opus 4.6 nor Sonnet 4.6 could solve. This positions it as a meaningful step forward for autonomous coding agents and complex engineering workflows.

3.Financials

Per-token price unchanged at $5/M input and $25/M output; prompt caching cuts costs up to 90%

Batch processing provides 50% cost savings. US-only inference is available at a 1.1x price premium on both input and output tokens.

4.Financials

New tokenizer generates 32–45% more native tokens for identical text; OpenRouter analysis finds 12–27% actual cost increases

While Anthropic's per-token pricing is unchanged, OpenRouter's empirical study of users migrating from Opus 4.6 to 4.7 found the new tokenizer produces substantially more tokens: ~45% more for inputs under 2K tokens, ~42% for 2K–10K, ~34% for 10K–25K, and ~32–33% for inputs above 25K. Prompt caching absorbs most inflation for large prompts (93% for 128K+), but short-prompt, non-cached workloads bear the full cost increase. Anthropic disclosed a 1.0–1.35x inflation range; OpenRouter's data shows up to 1.45x for the shortest prompts.

5.Insight

Short-prompt, low-cache workloads are most exposed to tokenizer-driven cost increases

For API users sending short, one-off prompts — chatbots, quick lookups, real-time queries — the cache offers little relief and the full token inflation applies. High-volume enterprise users with large, repetitive system prompts are better positioned to neutralize the impact through caching.

6.Security Alert

Opus 4.7 is the first model to carry Project Glasswing automated cybersecurity request blocking

Project Glasswing auto-blocks prohibited high-risk cybersecurity requests at inference time. Anthropic also launched a Cyber Verification Program to give legitimate security researchers, pen testers, and red-teamers a pathway to use the model's security capabilities.

7.Strategy

Project Glasswing sets a precedent: safety infrastructure as prerequisite for high-capability access

By coupling automated blocking and a verification program to Opus 4.7, Anthropic signals that its most powerful future models will require demonstrated safety controls before enabling sensitive use cases — framing safety as a product feature, not just a compliance obligation.

8.Infrastructure

Available on Claude API, Amazon Bedrock, Google Cloud Vertex AI, and Microsoft Foundry

Multi-cloud availability means enterprise customers can access Opus 4.7 within existing cloud procurement relationships and compliance frameworks. US-only inference routing available at 1.1x premium for data residency requirements.

Product Launch = new model/feature release, Stat = benchmark or quantitative result, Financials = pricing and cost structure, Insight = analytical inference from data, Security Alert = safety or security feature, Strategy = business positioning, Infrastructure = deployment and routing options

What This Means

Opus 4.7 is a meaningful step forward for teams building autonomous coding agents and long-horizon enterprise workflows: the 1M token context window and adaptive thinking together make it practical to delegate complex, multi-session engineering tasks with less human oversight. However, buyers should look past the unchanged per-token rates — according to OpenRouter's empirical analysis of real user traffic, the new tokenizer generates 32–45% more tokens for identical text, translating to effective cost increases of 12–27%. Prompt caching largely neutralizes this inflation for large, repetitive prompts, but short-prompt and low-cache workloads will see the full impact. The Project Glasswing debut also matters: Anthropic is explicitly coupling safety infrastructure to capability access, a pattern likely to define how its most powerful future models are gated.

Sentiment

Mostly excited about coding and agent improvements, with notable criticism on performance regressions and safety measures

@Thorium_LabsJoão Edgar Sallay · Founder & CEO | Agentic AI & Hyper Immersive ExperiencesView post
Excited

Opus 4.7 just landed. And early testers say you can finally hand off your hardest coding work without constant supervision... delivers a clear step up from Opus 4.6 in advanced software engineering.

@aidailyprimerAI Engineer Pigeon · Daily brief for AI EngineersView post
Mixed

opus 4.7 launch day report... SWE-Bench Verified +6.8 MRCR at 1M context: −46 (lol) new tokeniser makes every input 1.0–1.35x fatter claude code defaulted to xhigh so yes your rate limit IS getting eaten faster

@PatrickMoorheadPatrick Moorhead · Founder, CEO, Chief Analyst @MoorInsightsStratView post
Disappointed

Opus 4.7 is having a very hard time doing even basic things, like exporting work to a GDoc. Something Opus always could do. This is really, really bad.

@thesincerevpThe Sincere VP · Senior technical advisor at one of the Project Glasswing organizationsView post
Alarmed

Anthropic built their most powerful generally available model and then deliberately made it worse at one specific thing. Hacking... That's not a guardrail. That's surgery.

Insider view breaking down the deliberate reduction of cyber capabilities

Split

~70/30 positive/negative — developers impressed with autonomous coding and 1M context vs critics on regressions, costs, and gated cyber powers.

Sources

Updates

Apr 29

Added empirical tokenizer cost analysis from OpenRouter's study of users migrating from Opus 4.6 to 4.7. New data: tokenizer generates 32–45% more native tokens for identical text, translating to 12–27% effective cost increases in practice. Prompt caching absorbs 64–93% of inflation for large prompts (25K+) but short prompts bear full impact. Anthropic's disclosed 1.0–1.35x inflation range understated by OpenRouter's measurements (up to 1.45x for shortest prompts). Updated tier1_scan, tier2_understand, tier3_deep_dive, what_this_means, and key_facts accordingly.

Apr 17

Updated with materially new details from two additional sources: 1M token context window, adaptive thinking feature (auto-scales reasoning depth by task complexity), hybrid reasoning model classification, prompt caching (up to 90% cost savings), batch processing (50% cost savings), US-only inference at 1.1x pricing, and consumer tier availability (Pro, Max, Team, Enterprise). Key fact corrected: benchmark gain description now specifies "4 tasks that Opus 4.6 and Sonnet 4.6 could not solve" rather than "previously unsolvable" (per editor review). Importance score raised from 7 to 8; is_top_news set to true.

Similar Events