Anthropic launched Opus 4.7 on April 16, 2026, with the most reassuring formula possible: unchanged pricing, $5 per million tokens for input, $25 for output.
The Claude API bill sent thirty days later is 10 to 35% higher than expected for most teams that migrated without adjustments.
The explanation lies in a line from the migration guide: the new tokenizer splits the same text into more tokens, and Simon Willison measured 1.46x on a real system prompt.
The price per kilo hasn’t changed, it’s the scale that shows more on the merchant’s side, and this article outlines seven ways to compensate without sacrificing quality.
In short
- Measure before migrating: run your system prompt through Willison’s token counter to calibrate the real multiplier between 1.0x and 1.46x
- Route across three models: reserve Opus 4.7 for tough tasks, Sonnet 4.6 for 80% of traffic, Haiku 4.5 for classification, saving 55% on 100M monthly tokens
- Stack Batch API and cache: for nighttime enrichment of 50M tokens on Sonnet, bill drops from €150 to €7.50/month, saving €1,710 annually
- Frame Claude Code: replace the default xhigh tier with a mix of xhigh 30% and high 70%, saving €345 annually for ten developers
- Equip agents with task budgets: 20,000 tokens minimum per task to avoid the scenario of being billed €47,000 overnight by a loop
Unchanged displayed pricing, rising real bill: what Anthropic didn’t say
The Opus 4.7 pricing announced on April 16, 2026, mirrors that of Opus 4.6: $5 for a million input tokens, $25 for a million output tokens.
The official statement slips in a sentence that few teams read to the end: “the same text can generate 1.0x to 1.35x tokens depending on the content type”.
Simon Willison turned this discreet clause into a public figure four days later: his counter shows 7,335 tokens for Opus 4.7 on a system prompt that required 5,039 on Opus 4.6, a measured multiplier of 1.46x.
For an SME CTO who charges a monthly report at €99 to clients, the Claude cost per report shifts from €18 to €24, reducing the margin by six points on a recurring revenue line.
Compare this to a merchant who keeps the price per kilo but replaces their scale with one that shows 10 to 35% more on the same basket.
The quote of €300 for automating a client onboarding becomes a quote of €405 without any contract parameter change.
Quality rises, vision capabilities triple, and part of the inflation is recouped in more precise responses.
You can compare these quality promises with field measurements in our analysis on what Claude Opus 4.7 really changes in terms of quality.
Breaking down the numbers: 1.0x to 1.35x ratios depending on content
The multiplier announced by Anthropic hides a wide dispersion depending on what you run through the model.
Text, code, JSON, and French: ratios by content type
On a 30-page text PDF, Willison measures 1.08x (60,934 tokens versus 56,482).
On a system prompt filled with structured instructions, he measures 1.46x, above the official ceiling.
Independent tests converge on 1.25x for standard English and 1.30x to 1.35x for code, JSON, and Korean.
A team generating 80% structured JSON will see their Claude API bill climb towards the upper range.
High-resolution vision: 3.01x on 3.7 MP images, nothing on small ones
The most spectacular surge affects vision: Willison measures 3.01x on a PNG of 3,456 × 2,234 pixels.
On an image of 682 × 318 pixels, the counts are indistinguishable: 314 tokens for 4.7 versus 310 for 4.6.
The 3x only hits if you exploit the new limit of 2,576 pixels on the long side (3.75 megapixels).
Comparative pricing April 2026: Opus 4.7, Sonnet 4.6, Haiku 4.5
Official prices place Opus 4.7 at $5/25 per MTok, Sonnet 4.6 at $3/15, Haiku 4.5 at $1/5.
The Batch API applies a 50% discount on any model, cache reads a 0.1x factor, and cache writes 1.25x for a 5-minute TTL or 2.0x for 1 hour.
Levers 1 and 2: smart routing and switching to Batch API
The first two levers tackle the root of the expense: the chosen model and call mode.
Ternary model routing: Opus, Sonnet, Haiku based on difficulty
A lightweight router upstream that detects complexity and directs it to Haiku 4.5 for classification, Sonnet 4.6 for 80% of routine traffic, and Opus 4.7 for tough cases cuts an average 55% off a 100M monthly token bill.
The mechanism is coded in 30 lines of Python with a rule based on context length and business keywords.
The trap to avoid: don’t route requests requiring multi-step reasoning to Haiku, or retries will cancel out the savings.
A more detailed decision framework is in our article on Claude’s multi-tool agents.
Batch API: 50% off on asynchronous loads
The Batch API returns 50% of any Claude API bill for processes that tolerate a 24-hour delay.
Natural use cases include nighttime summaries, corpus tagging, SEO meta-description generation, and internal digests.
For a 50M token monthly enrichment on Sonnet 4.6, the bill drops from €150 to €75 before applying other levers.
Only trap: US peaks between 8 am and 2 pm EST where Batch queues rise, to be shifted to UTC+1 to meet the 24-hour deadline.
Levers 3 and 4: prompt caching and thinking tiers for the Claude API bill
The third and fourth levers are about call configuration, without changing the model.
Prompt caching: 90% on cache reads, trap of regressing TTL
On Opus 4.7, a cache read costs $0.50/MTok, ten times less than the base rate.
The cache write pays 1.25x for a 5-minute TTL or 2.0x for 1 hour.
Concrete case: a customer support app sends 30,000 tokens of system prompt and 5,000 tokens of question, 100 requests per five-minute window, dropping from €10.50 to €3.675 with cache enabled, an immediate 65% savings.
Critical trap: the TTL window silently regressed from 1 hour to 5 minutes on Claude Code in March 2026 (GitHub issue #46829), forcing unexpected cache rewrites.
Another trap: any dynamic timestamp in the system prompt invalidates the cache with each call.
Tiers xhigh, high, medium, low: dosing without losing quality
Opus 4.7 adds the xhigh tier between high and max, activated by default on Claude Code.
The default xhigh is like having the air conditioning on full blast in a new car: refreshing, but it drains the tank visibly on multi-turn loops.
A team of ten developers switching from systematic xhigh to a mix of xhigh 30% and high 70% saves €345 per year without measurable quality loss on refactoring.
Empirical rule: low and medium for classification, high for standard reasoning, xhigh reserved for complex engineering problems with multiple dependencies.
The /compact command in Claude Code clears conversational memory between topics and prevents internal reflection from stacking turn by turn.
Levers 5, 6, and 7: efficient tool calling, task budgets, combinations
The last three levers target agent architectures where the bill escalates fastest.
Programmatic Tool Calling and Tool Search: 37% to 85% reduction
Programmatic Tool Calling shifts tool calls into a code snippet that the model writes once instead of sending the whole schema each turn, cutting 37% of tokens on 5-tool pipelines.
Tool Search indexes tool definitions and only loads the relevant subset, cutting up to 85% on agents with 20 or more tools.
Both mechanisms combine, and their configuration fits into two arguments of the Anthropic SDK 0.52.
Task budgets beta: insurance against €47,000 infinite loops
The beta task_budgets imposes a token cap on an agent task, 20,000 tokens minimum, and cleanly cuts the loop when the budget is consumed.
The countdown acts as a kitchen timer for the agent: it sees it ticking and prefers to wrap up its response rather than being interrupted mid-reasoning.
This guarantee avoids the nightmare scenario of an autonomous agent looping all night on a parsing bug and delivering a Claude API bill of €47,000 at 8 am.
The syntax is straightforward in the Python SDK, and the cap is set per task.
Stacked Batch API and prompt caching: up to 95% on recurring enrichment
The levers stack when they target different tokens.
Example: nighttime enrichment of 50M tokens on Sonnet 4.6, with a stable system prompt at 90% and variable requests at 10%, dropping from €150/month to €7.50/month, a 94.95% savings and €1,710 annually.
Opus 4.6 to 4.7 migration plan in 4 phases, API traps to flag
The Opus 4.6 to 4.7 switch is managed in four phases that protect the Claude API bill and service continuity.
Phase 1: measure on 5 to 10 representative prompts with the official token counter to calibrate the real traffic multiplier.
Phase 2: 5% canary, redirect 5% of production traffic to Opus 4.7 for 48 hours comparing latency, quality, and cost.
Phase 3: progressive 5 / 25 / 50 / 100%, ramp up in stages with automatic rollback if the bill exceeds the planned budget by more than 20%.
Phase 4: lock-in, activate task budgets and prompt caching by default, and monitor 400 error returns for 72 hours.
The breaking APIs to flag: temperature, top_p, and top_k outside default values now return 400 on Opus 4.7.
The thinking display disappears from the default rendering and must be explicitly reactivated if your UX displays it.
The TTL regression from 1 hour to 5 minutes on Claude Code remains the most costly silent trap of 2026: a team that doesn’t set the value in configuration may see their bill double for no apparent reason.
Monitor cost anomaly returns with a circuit breaker at 2x the expected budget: it’s the assurance that a poorly formed prompt won’t burn a month’s budget in an hour.
To put these numbers in the context of the model’s new capabilities, reread the Opus 4.7 announcement on April 16, 2026.
Claude API bill inflation is a fact, but it remains manageable for any team that refuses blind migration.
The hierarchy by ROI and implementation effort is in three stages: cache and routing for most savings in two days, batch and thinking tiers for the next layer in two weeks, efficient tool calling and task budgets to lock in agent architectures over a quarter.
For a solo developer, cache switch plus Haiku on simple tasks is enough to absorb inflation.
For a multi-tenant SaaS SME, routing plus task budgets becomes the foundation for allocating costs per client.
For an autonomous agent team, task budgets plus tool search plus batch are non-negotiable.
The Claude API bill should no longer surprise: it should arrive to the euro where the budget placed it.
FAQ on the Claude API bill after Opus 4.7
Did Anthropic increase the price of Claude Opus 4.7?
No, the price per token remains at $5/MTok for input and $25/MTok for output, the extra cost comes from the tokenizer splitting the same text into 1.0 to 1.35 times more tokens depending on the content.
How much will my Claude API bill increase if I migrate to Opus 4.7 without changing anything?
Expect 10 to 35% on average, and up to 40% on structured system prompts measured by Simon Willison.
Is it better to stay on Opus 4.6 to save money?
Yes if your use is strictly textual and 4.6 quality meets your needs, no if your workflow benefits from high-resolution vision or xhigh reasoning, absent from 4.6.
How to choose between low, medium, high, and xhigh tiers?
Low and medium for classification and structured generation, high for standard reasoning, xhigh for complex engineering with multiple dependencies.
How to implement prompt caching in less than an hour?
In the Python SDK, add the cache_control type ephemeral parameter to the system block of the request, and explicitly set the TTL to 5 minutes or 1 hour depending on call frequency.
Which workflows to switch to Batch API without breaking the product?
Nighttime summaries, corpus tagging, meta-description generation, and internal digests tolerate the 24-hour delay, everything else must remain synchronous.
How to prevent an autonomous agent from blowing up my bill overnight?
Activate task_budgets with a minimum 20,000 token cap per task and add a circuit breaker on the orchestrator side at 2x the expected budget.
Is model routing between Haiku, Sonnet, and Opus worth the complexity?
Yes beyond 10M monthly tokens, as 55% savings on 100M tokens more than covers the 30 lines of router code.
How to allocate Claude costs per client in a multi-tenant application?
Instrument each call with a client tag, aggregate in your observability (Datadog, Finout, CloudZero) and monitor clients drawing more than 3x the average.
Does the TTL cache regression from 1 hour to 5 minutes affect the Claude API directly?
No, the issue only affected Claude Code in March 2026 (GitHub issue #46829), on the API side the TTL is explicitly set per call and remains under your control.
Related Articles
Building agents that don’t break: OpenAI’s new stack explained
By 2026, no digital company questions whether to integrate an AI agent somewhere: customer service, quote generation, ticket sorting, contract analysis. The real question has become: how to build them…
Claude Opus 4.7 design: will we still need graphic designers or front-end devs in 2026?
On April 16, 2026, Anthropic launched Claude Opus 4.7 design, a model with image resolution tripled and a SWE-bench Pro score jumping from 53.4% to 64.3% in just one version….